Harshay Shah (@harshays_) 's Twitter Profile
Harshay Shah

@harshays_

ML PhD student

ID: 591163833

linkhttp://harshay.me calendar_today26-05-2012 17:03:22

21 Tweet

533 Followers

691 Following

Harshay Shah (@harshays_) 's Twitter Profile Photo

Neural nets can generalize well on test data, but often lack robustness to distributional shifts & adversarial attacks. Our #NeurIPS2020 paper on simplicity bias sheds light on this phenomenon. Poster: session #4, town A2, spot C0, 12pm ET today! Paper: bit.ly/39RXDel

Neural nets can generalize well on test data, but often lack robustness to distributional shifts & adversarial attacks.

Our #NeurIPS2020 paper on simplicity bias sheds light on this phenomenon. 

Poster: session #4, town A2, spot C0, 12pm ET today!
Paper: bit.ly/39RXDel
Harshay Shah (@harshays_) 's Twitter Profile Photo

Do input gradients highlight discriminative and task-relevant features? Our #NeurIPS2021 paper takes a three-pronged approach to evaluate the fidelity of input gradient attributions. Poster: session 3, spot C0 Paper: bit.ly/3EzdvyH with Prateek Jain and @pnetrapalli

Do input gradients highlight discriminative and task-relevant features? 

Our #NeurIPS2021 paper takes a three-pronged approach to evaluate the fidelity of input gradient attributions.

Poster: session 3, spot C0 
Paper: bit.ly/3EzdvyH
with <a href="/jainprateek_/">Prateek Jain</a> and @pnetrapalli
Aleksander Madry (@aleks_madry) 's Twitter Profile Photo

You’re deploying an ML system, choosing between two models trained w/ diff algs. Same training data, same acc... how do you differentiate their behavior? ModelDiff (gradientscience.org/modeldiff) lets you compare *any* two learning algs! w/ Harshay Shah Sam Park Andrew Ilyas (1/8)

You’re deploying an ML system, choosing between two models trained w/ diff algs. Same training data, same acc... how do you differentiate their behavior?

ModelDiff (gradientscience.org/modeldiff) lets you compare *any* two learning algs!
w/ <a href="/harshays_/">Harshay Shah</a> <a href="/smsampark/">Sam Park</a> <a href="/andrew_ilyas/">Andrew Ilyas</a> (1/8)
Andrew Ilyas (@andrew_ilyas) 's Twitter Profile Photo

TRAK, our latest work on data attribution (trak.csail.mit.edu), speeds up datamodels up to 1000x! ➡️ our earlier work ModelDiff (w/ Harshay Shah Sam Park Aleksander Madry) can now compare any two learning algorithms in larger-scale settings. Try it out: github.com/MadryLab/model…

Harshay Shah (@harshays_) 's Twitter Profile Photo

If you are at #ICML2023 today, check out our work on ModelDiff, a model-agnostic framework for pinpointing differences between any two (supervised) learning algorithms! Poster: #407 at 2pm (Wednesday) Paper: icml.cc/virtual/2023/p… w/ Sam Park Andrew Ilyas Aleksander Madry

If you are at #ICML2023 today, check out our work on ModelDiff, a model-agnostic framework for pinpointing differences between any two (supervised) learning algorithms!

Poster: #407 at 2pm (Wednesday) 
Paper: icml.cc/virtual/2023/p…
w/ <a href="/smsampark/">Sam Park</a> <a href="/andrew_ilyas/">Andrew Ilyas</a> <a href="/aleks_madry/">Aleksander Madry</a>
Harshay Shah (@harshays_) 's Twitter Profile Photo

New work with Andrew Ilyas and Aleksander Madry on tracing predictions back to individual components (conv filters, attn heads) in the model! Paper: arxiv.org/abs/2404.11534 Thread: 👇

Aleksander Madry (@aleks_madry) 's Twitter Profile Photo

How is an LLM actually using the info given to it in its context? Is it misinterpreting anything or making things up? Introducing ContextCite: a simple method for attributing LLM responses back to the context: gradientscience.org/contextcite w/ Ben Cohen-Wang, Harshay Shah, Kristian Georgiev

MIT CSAIL (@mit_csail) 's Twitter Profile Photo

How do black-box neural networks transform raw data into predictions? Inside these models are thousands of simple "components" working together. New MIT CSAIL research (bit.ly/473lcfE) introduces a method that helps us understand how these components compose to affect

MIT CSAIL (@mit_csail) 's Twitter Profile Photo

How can we really know if a chatbot is giving a reliable answer? 🧵 MIT CSAIL’s "ContextCite" tool can ID the parts of external context used to generate any particular statement from a language model, improving trust by helping users easily verify the statement:

How can we really know if a chatbot is giving a reliable answer? 🧵

MIT CSAIL’s "ContextCite" tool can ID the parts of external context used to generate any particular statement from a language model, improving trust by helping users easily verify the statement:
Harshay Shah (@harshays_) 's Twitter Profile Photo

MoEs provide two knobs for scaling: model size (total params) + FLOPs-per-token (via active params). What’s the right scaling strategy? And how does it depend on the pretraining budget? Our work introduces sparsity-aware scaling laws for MoE LMs to tackle these questions! 🧵👇

Ben Cohen-Wang (@bcohenwang) 's Twitter Profile Photo

It can be helpful to pinpoint the in-context information that a language model uses when generating content (is it using provided documents? or its own intermediate thoughts?). We present Attribution with Attention (AT2), a method for doing so efficiently and reliably! (1/8)

It can be helpful to pinpoint the in-context information that a language model uses when generating content (is it using provided documents? or its own intermediate thoughts?). We present Attribution with Attention (AT2), a method for doing so efficiently and reliably! (1/8)
Harshay Shah (@harshays_) 's Twitter Profile Photo

If you’re at #ICLR2025, go watch Vimal Thilak🦉🐒 give an oral presentation at the @SparseLLMs workshop on scaling laws for pretraining MoE LMs! Had a great time co-leading this project with Samira Abnar & Vimal Thilak🦉🐒 at Apple MLR last summer. When: Sun Apr 27, 9:30a Where: Hall 4-07