Sarthak Mittal (@sarthmit) 's Twitter Profile
Sarthak Mittal

@sarthmit

Graduate Student at @Mila_Quebec and Visiting Researcher @Meta. Prior Research Intern at @Apple, @MorganStanley, @NVIDIAAI and @YorkUniversity

ID: 1092905072026103809

linkhttps://sarthmit.github.io/ calendar_today05-02-2019 21:57:29

214 Tweet

645 Followers

732 Following

Mohammad Pezeshki (@mpezeshki91) 's Twitter Profile Photo

If you're interested in how to keep challenging neural networks throughout training, check out our latest preprint! #sample_efficiency #scaling_laws

If you're interested in how to keep challenging neural networks throughout training, check out our latest preprint!
#sample_efficiency #scaling_laws
Neel Nanda (@neelnanda5) 's Twitter Profile Photo

GDM Mech Interp Update: We study if SAEs help probes generalise OOD (they don't 😢). Based on this + parallel negative results on real-world tasks, we're de-prioritising SAE work. Our guess is that SAEs aren't useless, but also aren't a game-changer More + new research in 🧵

GDM Mech Interp Update:

We study if SAEs help probes generalise OOD (they don't 😢). Based on this + parallel negative results on real-world tasks, we're de-prioritising SAE work.

Our guess is that SAEs aren't useless, but also aren't a game-changer

More + new research in 🧵
Alex Tong (@alexandertong7) 's Twitter Profile Photo

FPI workshop off to a great start with Emtiyaz Khan talking about Adaptive Bayesian Intelligence! Come check it out in Peridot 202-203 #FPIWorkshop #ICLR25

FPI workshop off to a great start with Emtiyaz Khan talking about Adaptive Bayesian Intelligence! Come check it out in Peridot 202-203 #FPIWorkshop #ICLR25
Sarthak Mittal (@sarthmit) 's Twitter Profile Photo

Come check out the workshop and hear about novel works and contributions from an exciting lineup of speakers and panelists!

Alex Tong (@alexandertong7) 's Twitter Profile Photo

Great talk by Grant Rotskoff linking sampling with nonequilibrium physics in the #FPIworkshop. Come by for the poster session (Peridot 202-203) for the next hour. #ICLR25

Great talk by Grant Rotskoff linking sampling with nonequilibrium physics in the #FPIworkshop. Come by for the poster session (Peridot 202-203) for the next hour.  #ICLR25
Divyat Mahajan (@divyat09) 's Twitter Profile Photo

Happy to share that Compositional Risk Minimization has been accepted at #ICML2025 📌Extensive theoretical analysis along with a practical approach for extrapolating classifiers to novel compositions! 📜 arxiv.org/abs/2410.06303

Happy to share that Compositional Risk Minimization has been accepted at #ICML2025

📌Extensive theoretical analysis along with a practical approach for extrapolating classifiers to novel compositions!

📜 arxiv.org/abs/2410.06303
Mandana Samiei (@mandanasamiei) 's Twitter Profile Photo

A great collab with former labmates Anthony GX-Chen & Dongyan! Interesting cognitive limitation in LMs: strong disjunctive bias leads to poor performance on conjunctive causal inference tasks. Mirrors adult human biases—possibly a byproduct of training data priors.

Andrea Dittadi (@andrea_dittadi) 's Twitter Profile Photo

Physics says it's fine to be lazy New preprint on minimum-excess-work guidance: arxiv.org/abs/2505.13375 Check out the thread below 👇

Arnav Jain (@arnavkj95) 's Twitter Profile Photo

⛵️ Excited to share 𝚂𝙰𝙸𝙻𝙾𝚁: a method for *learning to search* with learned world + reward models to plan in the latent space at test-time. Unlike behavior cloning, 𝚂𝙰𝙸𝙻𝙾𝚁 recovers from mistakes without any additional data, DAgger corrections, or ground truth rewards.

Joey Bose (@bose_joey) 's Twitter Profile Photo

🎉Personal update: I'm thrilled to announce that I'm joining Imperial College London Imperial College London as an Assistant Professor of Computing Imperial Computing starting January 2026. My future lab and I will continue to work on building better Generative Models 🤖, the hardest

Sarthak Mittal (@sarthmit) 's Twitter Profile Photo

Explicit latents or implicit marginalization? at #ICML2025 📌 Tue, 11 am 📍East Exhibition Hall A-B (E-1603) Come check out surprising results on whether explicitly incentivizing learning of correct latents improves generalization over implicitly marginalizing it!

Explicit latents or implicit marginalization? at #ICML2025 

📌 Tue,  11 am 
📍East Exhibition Hall A-B (E-1603)

Come check out surprising results on whether explicitly incentivizing learning of correct latents improves generalization over implicitly marginalizing it!