Ian Tenney (@iftenney@sigmoid.social) (@iftenney) 's Twitter Profile
Ian Tenney (@[email protected])

@iftenney

Staff Research Scientist, People + AI Research @GoogleAI #GoogleResearch. Interpretability, analysis, and visualizations for LLMs. Opinions my own.

ID: 898538970

linkhttp://iftenney.github.io calendar_today22-10-2012 22:16:38

92 Tweet

1,1K Followers

530 Following

Google AI (@googleai) 's Twitter Profile Photo

Recent innovation has given rise to #ML models w/ impressive capabilities, but there’s much to learn about how we attribute model behavior to training data, algorithms, architecture, & more! Have papers or ideas on this? Submit to ATTRIB @ #NeurIPS2023 → attrib-workshop.cc

Recent innovation has given rise to #ML models w/ impressive capabilities, but there’s much to learn about how we attribute model behavior to training data, algorithms, architecture, & more! Have papers or ideas on this? Submit to ATTRIB @ #NeurIPS2023 → attrib-workshop.cc
iislucas (Lucas Dixon) (@iislucas) 's Twitter Profile Photo

PAIR is looking for a Research Scientist interested in making hard ML problems (like understanding langauge) much smaller... In Paris, and working closely with fun interactive explorable visualizations too. See: goo.gle/3PcvPEs

Jeff Dean (@jeffdean) 's Twitter Profile Photo

We're also releasing a Responsible Generative AI Toolkit that provides resources to apply best practices for responsible use of open models such as the Gemma models, including: Guidance on setting safety policies, safety tuning, safety classifiers and model evaluation. The

Ian Tenney (@iftenney@sigmoid.social) (@iftenney) 's Twitter Profile Photo

Super excited for the Gemma model release, and with it a new debugging tool we built on 🔥LIT - use gradient-based salience to debug and refine complex LLM prompts! ai.google.dev/responsible/mo…

Ian Tenney (@iftenney@sigmoid.social) (@iftenney) 's Twitter Profile Photo

Very, very cool work from Asma Ghandeharioun and colleagues - Patchscopes is a super flexible way to decode knowledge from internal states of an LLM, bridging between interpretability, causality, and control.

Ian Tenney (@iftenney@sigmoid.social) (@iftenney) 's Twitter Profile Photo

New open-source tool: LLM Comparator can help make sense of side-by-side model evals. In-browser demo at pair-code.github.io/llm-comparator/, and read more in the blog post below.

Ian Tenney (@iftenney@sigmoid.social) (@iftenney) 's Twitter Profile Photo

Try out LLM Comparator to help make sense of LLM evaluations! Alongside the Gemma 2 launch, we've released a Python library to help run rater models, bulletizing, and clustering - so all you need are prompts and model responses. Scripts and notebooks at goo.gle/llm-comparator

Ian Tenney (@iftenney@sigmoid.social) (@iftenney) 's Twitter Profile Photo

Check out our new work on scaling training data attribution (TDA) toward LLM pretraining - and some interesting things we found along the way. arxiv.org/abs/2410.17413 and more below from most excellent student researcher Tyler Chang ⬇️

Tyler Chang (@tylerachang) 's Twitter Profile Photo

Presenting our work on training data attribution for pretraining this morning: iclr.cc/virtual/2025/p… -- come stop by in Hall 2/3 #526 if you're here at ICLR!

Ian Tenney (@iftenney@sigmoid.social) (@iftenney) 's Twitter Profile Photo

Submission deadline extended to May 19! Working on or have thoughts on real-world applications of interpretability, and how we can use it in practice? Consider submitting to our workshop at ICML 2025. More at actionable-interpretability.github.io and below👇

Mor Geva (@megamor2) 's Twitter Profile Photo

Going to #icml2025? Don't miss the Actionable Interpretability Workshop (Actionable Interpretability Workshop ICML2025)! We've got an amazing lineup of speakers, panelists, and papers, all focused on leveraging insights from interpretability research to tackle practical, real-world problems ✨

Going to #icml2025? Don't miss the Actionable Interpretability Workshop (<a href="/ActInterp/">Actionable Interpretability Workshop ICML2025</a>)! We've got an amazing lineup of speakers, panelists, and papers, all focused on leveraging insights from interpretability research to tackle practical, real-world problems ✨
Tal Haklay (@tal_haklay) 's Twitter Profile Photo

🚨Meet our panelists at the Actionable Interpretability Workshop Actionable Interpretability Workshop ICML2025 at ICML Conference! Join us July 19 at 4pm for a panel on making interpretability research actionable, its challenges, and how the community can drive greater impact. Naomi Saphra hiring my lab at ICML 🧈🪰 Samuel Marks Kyle Lo Fazl Barez

🚨Meet our panelists at the Actionable Interpretability Workshop <a href="/ActInterp/">Actionable Interpretability Workshop ICML2025</a> at <a href="/icmlconf/">ICML Conference</a>!

Join us July 19 at 4pm for a panel on making interpretability research actionable, its challenges, and how the community can drive greater impact.
<a href="/nsaphra/">Naomi Saphra hiring my lab at ICML 🧈🪰</a> <a href="/saprmarks/">Samuel Marks</a> <a href="/kylelostat/">Kyle Lo</a> <a href="/FazlBarez/">Fazl Barez</a>