Sam Park (@smsampark) 's Twitter Profile
Sam Park

@smsampark

Postdoc at Stanford CS. Previously: MIT PhD, Cornell BS

ID: 1607751613

linkhttps://sungminpark.com calendar_today20-07-2013 07:54:39

53 Tweet

403 Followers

476 Following

Sara Hooker (@sarahookr) 's Twitter Profile Photo

I gave a keynote this week at the fantastic ATTRIB Workshop #NeurIPS2023 "What does scale give us: Why we are building a ladder đŸȘœ to the moon 🌕" Some of you asked for my slides, sharing below: docs.google.com/presentation/d
 Thanks to the organizers for a fantastic workshop! đŸ”„

Aleksander Madry (@aleks_madry) 's Twitter Profile Photo

How do we attribute an image generated by a diffusion model back to the training data? w/ Kristian Georgiev Josh Vendrow Hadi Salman Sam Park we show that it’s useful to look at each step of the diffusion process:

Aleksander Madry (@aleks_madry) 's Twitter Profile Photo

We tend to choose LM training data via intuitive notions of text quality... but LMs are often *un*intuitive. Is there a better way? w/Logan Engstrom, Axel Feldmann: we select better data by modeling how models learn from data. Our method, DsDm, can greatly improve

We tend to choose LM training data via intuitive notions of text quality... but LMs are often *un*intuitive. Is there a better way?

w/<a href="/logan_engstrom/">Logan Engstrom</a>, <a href="/axel_s_feldmann/">Axel Feldmann</a>: we select better data by modeling how models learn from data. Our method, DsDm, can greatly improve
Aleksander Madry (@aleks_madry) 's Twitter Profile Photo

Our second (and final) blog post on model components is now out: we show that component attributions enable model editing! See the blog post for more: gradientscience.org/modelcomponent
 Paper: arxiv.org/abs/2404.11534 (this time with no typos :) Code: github.com/MadryLab/model


Sarah Cen (@cen_sarah) 's Twitter Profile Photo

In work w/ Andrew Ilyas Jennifer Allen Hannah Li Aleksander Madry we give experimental evidence that users strategize on recommender systems! We find that users react to their (beliefs about) *algorithms* (not just content!) to shape future recs. Paper: arxiv.org/abs/2405.05596 1/8

In work w/ <a href="/andrew_ilyas/">Andrew Ilyas</a> <a href="/_JenAllen/">Jennifer Allen</a> <a href="/hannahq_li/">Hannah Li</a>  <a href="/aleks_madry/">Aleksander Madry</a> we give experimental evidence that users strategize on recommender systems!

We find that users react to their (beliefs about) *algorithms* (not just content!) to shape future recs.

Paper: arxiv.org/abs/2405.05596

1/8
Aleksander Madry (@aleks_madry) 's Twitter Profile Photo

In ML, we train on biased (huge) datasets âžĄïž models encode spurious corrs and fail on minority groups. Can we scalably remove "bad" data? w/ Saachi Jain Kimia Hamidieh Kristian Georgiev Andrew Ilyas Marzyeh we propose D3M, a method for exactly this: gradientscience.org/d3m/

In ML, we train on biased (huge) datasets âžĄïž models encode spurious corrs and fail on minority groups. Can we scalably remove "bad" data?

w/ <a href="/saachi_jain_/">Saachi Jain</a> <a href="/kimiahmdh/">Kimia Hamidieh</a> <a href="/kris_georgiev1/">Kristian Georgiev</a> <a href="/andrew_ilyas/">Andrew Ilyas</a> <a href="/MarzyehGhassemi/">Marzyeh</a> we propose D3M, a method for exactly this: gradientscience.org/d3m/
Aleksander Madry (@aleks_madry) 's Twitter Profile Photo

At #ICML2024 ? Our tutorial "Data Attribution at Scale" will be to tomorrow at 9:30 AM CEST in Hall A1! I will not be able to make it (but will arrive later that day), but my awesome students Andrew Ilyas Sam Park Logan Engstrom will carry the torch :)

Aleksander Madry (@aleks_madry) 's Twitter Profile Photo

Attending #ICML2024? Check out our work on decomposing predictions and editing model behavior via targeted interventions to model internals! Poster: #2513, Hall C 4-9, 1:30p (Tue) Paper: arxiv.org/abs/2404.11534 w/ Harshay Shah Andrew Ilyas

Attending #ICML2024? Check out our work on decomposing predictions and editing model behavior via targeted interventions to model internals!

Poster: #2513, Hall C 4-9, 1:30p (Tue)  
Paper: arxiv.org/abs/2404.11534
w/ <a href="/harshays_/">Harshay Shah</a>  <a href="/andrew_ilyas/">Andrew Ilyas</a>
Logan Engstrom (@logan_engstrom) 's Twitter Profile Photo

Stop by our poster on model-aware dataset selection at ICML! Location/time: 1:30pm Hall C 4-9 #1010 (Tuesday) Paper: arxiv.org/abs/2401.12926 with: Axel Feldmann Aleksander Madry

Andrew Ilyas (@andrew_ilyas) 's Twitter Profile Photo

Thanks to all who attended our tutorial "Data Attribution at Scale" at ICML (w/ Sam Park Logan Engstrom Kristian Georgiev Aleksander Madry)! We're really excited to see the response to this emerging topic. Slides, notes, ICML video: ml-data-tutorial.org Public recording soon!

Thanks to all who attended our tutorial "Data Attribution at Scale" at ICML (w/ <a href="/smsampark/">Sam Park</a> <a href="/logan_engstrom/">Logan Engstrom</a> <a href="/kris_georgiev1/">Kristian Georgiev</a> <a href="/aleks_madry/">Aleksander Madry</a>)! We're really excited to see the response to this emerging topic.

Slides, notes, ICML video: ml-data-tutorial.org
Public recording soon!
Andrew Ilyas (@andrew_ilyas) 's Twitter Profile Photo

The ATTRIB workshop is back @ NeurIPS 2024! We welcome papers connecting model behavior to data, algorithms, parameters, scale, or anything else. Submit by Sep 18! More info: attrib-workshop.cc Co-organizers: Tolga Bolukbasi Logan Engstrom Sadhika Malladi Elisa Nguyen Sam Park

The ATTRIB workshop is back @ NeurIPS 2024! We welcome papers connecting model behavior to data, algorithms, parameters, scale, or anything else. Submit by Sep 18!

More info: attrib-workshop.cc

Co-organizers: <a href="/tolgab0/">Tolga Bolukbasi</a> <a href="/logan_engstrom/">Logan Engstrom</a> <a href="/SadhikaMalladi/">Sadhika Malladi</a> <a href="/_elinguyen/">Elisa Nguyen</a> <a href="/smsampark/">Sam Park</a>
Andrew Ilyas (@andrew_ilyas) 's Twitter Profile Photo

Have a paper on understanding how ML design choices relate to model behavior? A reminder that the (first) ATTRIB deadline is this Wednesday AoE! The second deadline is October 4th AoE, but one author will have to volunteer as an emergency reviewer.

Andrew Ilyas (@andrew_ilyas) 's Twitter Profile Photo

Machine unlearning ("removing" training data from a trained ML model) is a hard, important problem. Datamodel Matching (DMM): a new unlearning paradigm with strong empirical performance! w/ Kristian Georgiev Roy Rinberg Sam Park Shivam Garg Aleksander Madry Seth Neel (1/4)

Ken Liu (@kenziyuliu) 's Twitter Profile Photo

An LLM generates an article verbatim—did it “train on” the article? It’s complicated: under n-gram definitions of train-set inclusion, LLMs can complete “unseen” texts—both after data deletion and adding “gibberish” data. Our results impact unlearning, MIAs & data transparencyđŸ§”

An LLM generates an article verbatim—did it “train on” the article?

It’s complicated: under n-gram definitions of train-set inclusion, LLMs can complete “unseen” texts—both after data deletion and adding “gibberish” data. Our results impact unlearning, MIAs &amp; data transparencyđŸ§”
Zitong Yang (@zitongyang0) 's Twitter Profile Photo

To stress-test our technique, we apply Synthetic Continued Pretraining on 1K ICLR papers (joint effort with CLS), obtaining an LM that “knows” your ICLR submission in its weights. If you’d like to try out our LM, come to the oral or poster session below! Oral Session

To stress-test our technique, we apply Synthetic Continued Pretraining on 1K ICLR papers (joint effort with <a href="/ChengleiSi/">CLS</a>), obtaining an LM that “knows” your ICLR submission in its weights.

If you’d like to try out our LM, come to the oral or poster session below!

Oral Session
Tristan Thrush (@tristanthrush) 's Twitter Profile Photo

At #ICLR, check out Perplexity Correlations: a statistical framework to select the best pretraining data with no LLM training! I can’t make the trip, but Tatsunori Hashimoto will present the poster for us! Continue reading for the latest empirical validations of PPL Correlations:

Sam Park (@smsampark) 's Twitter Profile Photo

Cool application of training data attribution / influence estimation (using TRAK) to data selection for robot imitation learning!