Sleeper Agents: Training Deceptive LLMs that Persist Through Safety Training Paper • 2401.05566 • Published Jan 10 • 26
Sleeper Agents: Training Deceptive LLMs that Persist Through Safety Training Paper • 2401.05566 • Published Jan 10 • 26
Specific versus General Principles for Constitutional AI Paper • 2310.13798 • Published Oct 20, 2023 • 2
Measuring Faithfulness in Chain-of-Thought Reasoning Paper • 2307.13702 • Published Jul 17, 2023 • 27
Question Decomposition Improves the Faithfulness of Model-Generated Reasoning Paper • 2307.11768 • Published Jul 17, 2023 • 12
Towards Measuring the Representation of Subjective Global Opinions in Language Models Paper • 2306.16388 • Published Jun 28, 2023 • 6
Towards Measuring the Representation of Subjective Global Opinions in Language Models Paper • 2306.16388 • Published Jun 28, 2023 • 6
Opportunities and Risks of LLMs for Scalable Deliberation with Polis Paper • 2306.11932 • Published Jun 20, 2023 • 6
Language Models Don't Always Say What They Think: Unfaithful Explanations in Chain-of-Thought Prompting Paper • 2305.04388 • Published May 7, 2023 • 1