Staff Research Scientist, People + AI Research @GoogleAI #GoogleResearch. Interpretability, analysis, and visualizations for LLMs. Opinions my own.
ID: 898538970
http://iftenney.github.io 22-10-2012 22:16:38
92 Tweet
1,1K Followers
530 Following





Very, very cool work from Asma Ghandeharioun and colleagues - Patchscopes is a super flexible way to decode knowledge from internal states of an LLM, bridging between interpretability, causality, and control.



Check out our new work on scaling training data attribution (TDA) toward LLM pretraining - and some interesting things we found along the way. arxiv.org/abs/2410.17413 and more below from most excellent student researcher Tyler Chang ⬇️



Going to #icml2025? Don't miss the Actionable Interpretability Workshop (Actionable Interpretability Workshop ICML2025)! We've got an amazing lineup of speakers, panelists, and papers, all focused on leveraging insights from interpretability research to tackle practical, real-world problems ✨


🚨Meet our panelists at the Actionable Interpretability Workshop Actionable Interpretability Workshop ICML2025 at ICML Conference! Join us July 19 at 4pm for a panel on making interpretability research actionable, its challenges, and how the community can drive greater impact. Naomi Saphra hiring my lab at ICML 🧈🪰 Samuel Marks Kyle Lo Fazl Barez
