Daniela Amodei (@danielaamodei) 's Twitter Profile
Daniela Amodei

@danielaamodei

President @AnthropicAI. Formerly @OpenAI, @Stripe, congressional staffer, global development

ID: 373531234

linkhttp://anthropic.com calendar_today14-09-2011 19:02:13

29 Tweet

8,8K Followers

289 Following

Anthropic (@anthropicai) 's Twitter Profile Photo

Our first AI alignment paper, focused on simple baselines and investigations: A General Language Assistant as a Laboratory for Alignment arxiv.org/abs/2112.00861

Anthropic (@anthropicai) 's Twitter Profile Photo

Our first interpretability paper explores a mathematical framework for trying to reverse engineer transformer language models: A Mathematical Framework for Transformer Circuits: transformer-circuits.pub/2021/framework…

Anthropic (@anthropicai) 's Twitter Profile Photo

Our first societal impacts paper explores the technical traits of large generative models and the motivations and challenges people face in building and deploying them: arxiv.org/abs/2202.07785

Anthropic (@anthropicai) 's Twitter Profile Photo

In our second interpretability paper, we revisit “induction heads”. In 2+ layer transformers these pattern-completion heads form exactly when in-context learning abruptly improves. Are they responsible for most in-context learning in large transformers? transformer-circuits.pub/2022/in-contex…

Anthropic (@anthropicai) 's Twitter Profile Photo

On the @FLIxrisk podcast, we discuss AI research, AI safety, and what it was like starting Anthropic during COVID. futureoflife.org/2022/03/04/dan…

Anthropic (@anthropicai) 's Twitter Profile Photo

We've trained a natural language assistant to be more helpful and harmless by using reinforcement learning with human feedback (RLHF). arxiv.org/abs/2204.05862

We've trained a natural language assistant to be more helpful and harmless by using reinforcement learning with human feedback (RLHF). arxiv.org/abs/2204.05862
Anthropic (@anthropicai) 's Twitter Profile Photo

In a new paper, we show that repeating only a small fraction of the data used to train a language model (albeit many times) can damage performance significantly, and we observe a "double descent" phenomenon associated with this. arxiv.org/abs/2205.10487

In a new paper, we show that repeating only a small fraction of the data used to train a language model (albeit many times) can damage performance significantly, and we observe a "double descent" phenomenon associated with this.
arxiv.org/abs/2205.10487
Anthropic (@anthropicai) 's Twitter Profile Photo

Transformer MLP neurons are challenging to understand. We find that using a different activation function (Softmax Linear Units or SoLU) increases the fraction of neurons that appear to respond to understandable features without any performance penalty. transformer-circuits.pub/2022/solu/inde…

Transformer MLP neurons are challenging to understand.

We find that using a different activation function (Softmax Linear Units or SoLU) increases the fraction of neurons that appear to respond to understandable features without any performance penalty.

transformer-circuits.pub/2022/solu/inde…
Anthropic (@anthropicai) 's Twitter Profile Photo

In "Language Models (Mostly) Know What They Know", we show that language models can evaluate whether what they say is true, and predict ahead of time whether they'll be able to answer questions correctly. arxiv.org/abs/2207.05221

In "Language Models (Mostly) Know What They Know", we show that language models can evaluate whether what they say is true, and predict ahead of time whether they'll be able to answer questions correctly. arxiv.org/abs/2207.05221
Anthropic (@anthropicai) 's Twitter Profile Photo

Neural networks often pack many unrelated concepts into a single neuron – a puzzling phenomenon known as 'polysemanticity' which makes interpretability much more challenging. In our latest work, we build toy models where the origins of polysemanticity can be fully understood.

Anthropic (@anthropicai) 's Twitter Profile Photo

Introducing Claude 2! Our latest model has improved performance in coding, math and reasoning. It can produce longer responses, and is available in a new public-facing beta website at claude.ai in the US and UK.

Introducing Claude 2! Our latest model has improved performance in coding, math and reasoning. It can produce longer responses, and is available in a new public-facing beta website at claude.ai in the US and UK.