Suraj Srinivas (@suuraj) 's Twitter Profile
Suraj Srinivas

@suuraj

ml researcher at bosch ai / trying to understand why deep learning works

ID: 45312366

linkhttps://suraj-srinivas.github.io/ calendar_today07-06-2009 09:22:42

941 Tweet

1,1K Followers

1,1K Following

jack morris (@jxmnop) 's Twitter Profile Photo

# A new type of information theory this paper is not super well-known but has changed my opinion of how deep learning works more than almost anything else it says that we should measure the amount of information available in some representation based on how *extractable* it is,

# A new type of information theory

this paper is not super well-known but has changed my opinion of how deep learning works more than almost anything else

it says that we should measure the amount of information available in some representation based on how *extractable* it is,
Naomi Saphra hiring a lab 🧈🪰 (@nsaphra) 's Twitter Profile Photo

Take a break from arxiv/LW/AF. Sit in the woods with a random textbook and mull new ideas away from interp community lockstep. Diverge. Don’t compete with a saturated subtopic, maybe you’ll get to take weekends off. Premature overinvestment comes from monoculture.

☉rthonormalist🧭✡️ (@orthonormalist) 's Twitter Profile Photo

DID I CRACK IT? I think I figured out at least a chunk of the math. It's trade deficit divided by their exports. EU: exports 531.6, imports 333.4, deficit 198.2. 198.2/531.6 is 37, close to 39. Israel: exports 22.2, imports 14.8, deficit 7.4. 7.4/22.2 is 33.

Michal Moshkovitz (@ml_theorist) 's Twitter Profile Photo

In April 2024, we launched the Theory of Interpretable XAI seminar, aiming to build a community—unsure if we’d even have enough speakers. A year later, we’re still growing. New to the seminar? Join us in building the foundations of XAI together Tim van Erven Suraj Srinivas @ ICML 1/n

In April 2024, we launched the Theory of Interpretable XAI seminar, aiming to build a community—unsure if we’d even have enough speakers. A year later, we’re still growing.

New to the seminar? Join us in building the foundations of XAI together

<a href="/tverven/">Tim van Erven</a> <a href="/Suuraj/">Suraj Srinivas @ ICML</a> 

1/n
Michal Moshkovitz (@ml_theorist) 's Twitter Profile Photo

⏰⏰Theory of Interpretable AI Seminar ⏰⏰ Interested in Feature Attribution Explanations? In two weeks, May 6, Gunnar König Gunnar König will talk about "Disentangling Interactions and Dependencies in Feature Attribution" Tim van Erven Suraj Srinivas @ ICML

⏰⏰Theory of Interpretable AI Seminar ⏰⏰

Interested in Feature Attribution Explanations?

In two weeks, May 6, Gunnar König  <a href="/gcskoenig/">Gunnar König</a>  will talk about "Disentangling Interactions and Dependencies in Feature Attribution"

<a href="/tverven/">Tim van Erven</a> <a href="/Suuraj/">Suraj Srinivas @ ICML</a>
Michal Moshkovitz (@ml_theorist) 's Twitter Profile Photo

Curious about feature attribution? SHAP & LIME treat features independently—but features interact! Come hear how to "Disentangle Interactions and Dependencies in Feature Attribution" Tuesday (tomorrow!) 4pm CET, 10am ET Suraj Srinivas @ ICML Tim van Erven

arlo_son (@gson_ai) 's Twitter Profile Photo

#NLProc AI Co-Scientists 🤖 can generate ideas, but can they spot mistakes? (not yet! 🚫) In my recent paper, we introduce SPOT, a dataset of STEM manuscripts (math, materials science, chemistry, physics, etc), annotated with real errors. SOTA models like o3, gemini-2.5-pro

#NLProc
AI Co-Scientists 🤖 can generate ideas, but can they spot mistakes? (not yet! 🚫)

In my recent paper, we introduce SPOT, a dataset of STEM manuscripts (math, materials science, chemistry, physics, etc), annotated with real errors.

SOTA models like o3, gemini-2.5-pro
Michal Moshkovitz (@ml_theorist) 's Twitter Profile Photo

⏰⏰ Theory of Interpretable AI Seminar ⏰⏰ Chain-of-Thought: Why does explaining to LLMs using CoT prompting work? Join us on June 3, when Bohang Zhang @ICLR 2024 will dive into the mechanisms behind chain-of-thought prompting — and what makes it so effective Tim van Erven Suraj Srinivas @ ICML

⏰⏰ Theory of Interpretable AI Seminar ⏰⏰
Chain-of-Thought: Why does explaining to LLMs using CoT prompting work?

Join us on June 3, when <a href="/bohang_zhang/">Bohang Zhang @ICLR 2024</a> will dive into the mechanisms behind chain-of-thought prompting — and what makes it so effective

<a href="/tverven/">Tim van Erven</a>
<a href="/Suuraj/">Suraj Srinivas @ ICML</a>
Goodfire (@goodfireai) 's Twitter Profile Photo

We created a canvas that plugs into an image model’s brain. You can use it to generate images in real-time by painting with the latent concepts the model has learned. Try out Paint with Ember for yourself 👇

jack morris (@jxmnop) 's Twitter Profile Photo

## The case for more ambition i wrote about how AI researchers should ask bigger and simpler questions, and publish fewer papers:

## The case for more ambition

i wrote about how AI researchers should ask bigger and simpler questions,  and publish fewer papers:
Sebastian Bordt (@sbordt) 's Twitter Profile Photo

Have you ever wondered whether a few times of data contamination really lead to benchmark overfitting?🤔 Then our latest paper about the effect of data contamination on LLM evals might be for you!🚀 "How Much Can We Forget about Data Contamination?" (accepted at #ICML2025) shows

Michael Black (@michael_j_black) 's Twitter Profile Photo

Here's how my recent papers & reviews are going: * To solve a vision problem today, the sensible thing is to leverage a pre-trained VLM or video diffusion model. Such models implicitly represent a tremendous amount about the visual world that we can exploit. * Figure out how to