
Logan Engstrom
@logan_engstrom
CS PhD student @ MIT
ID: 1864965229
http://loganengstrom.com 14-09-2013 20:44:36
316 Tweet
1,1K Followers
331 Following

How is an LLM actually using the info given to it in its context? Is it misinterpreting anything or making things up? Introducing ContextCite: a simple method for attributing LLM responses back to the context: gradientscience.org/contextcite w/ Ben Cohen-Wang, Harshay Shah, Kristian Georgiev

In work w/ Andrew Ilyas Jennifer Allen Hannah Li Aleksander Madry we give experimental evidence that users strategize on recommender systems! We find that users react to their (beliefs about) *algorithms* (not just content!) to shape future recs. Paper: arxiv.org/abs/2405.05596 1/8


At #ICML2024 ? Our tutorial "Data Attribution at Scale" will be to tomorrow at 9:30 AM CEST in Hall A1! I will not be able to make it (but will arrive later that day), but my awesome students Andrew Ilyas Sam Park Logan Engstrom will carry the torch :)

Stop by our poster on model-aware dataset selection at ICML! Location/time: 1:30pm Hall C 4-9 #1010 (Tuesday) Paper: arxiv.org/abs/2401.12926 with: Axel Feldmann Aleksander Madry

Thanks to all who attended our tutorial "Data Attribution at Scale" at ICML (w/ Sam Park Logan Engstrom Kristian Georgiev Aleksander Madry)! We're really excited to see the response to this emerging topic. Slides, notes, ICML video: ml-data-tutorial.org Public recording soon!


The ATTRIB workshop is back @ NeurIPS 2024! We welcome papers connecting model behavior to data, algorithms, parameters, scale, or anything else. Submit by Sep 18! More info: attrib-workshop.cc Co-organizers: Tolga Bolukbasi Logan Engstrom Sadhika Malladi Elisa Nguyen Sam Park



After some very fun years at MIT, I'm really excited to be joining CMU as an assistant professor in Jan 2026! A big (huge!) thanks to my advisors (Aleksander Madry Constantinos Daskalakis), collaborators, mentors & friends. In the meantime, I'll be a Stein Fellow at Stanford Statistics.


Had a great time at Simons Institute for the Theory of Computing talking about new & upcoming work on meta-optimization of ML training tl;dr we show how to compute gradients *through* the training process & use them to optimize training. Immediate big gains on data selection, poisoning, attribution & more!

โHow will my model behave if I change the training data?โ Recent(-ish) work w/ Logan Engstrom: we nearly *perfectly* predict ML model behavior as a function of training data, saturating benchmarks for this problem (called โdata attributionโ).
