Virginia Smith (@gingsmith) 's Twitter Profile
Virginia Smith

@gingsmith

ML Professor @ CMU
🦋 gingsmith

ID: 451045487

linkhttps://www.cs.cmu.edu/~smithv/ calendar_today31-12-2011 00:00:49

31 Tweet

660 Followers

21 Following

ML@CMU (@mlcmublog) 's Twitter Profile Photo

blog.ml.cmu.edu/2025/01/08/opt… How can we train LLMs to solve complex challenges beyond just data scaling? In a new blogpost, Amrith Setlur, Yuxiao Qu Matthew Yang, Lunjun Zhang , Virginia Smith  and Aviral Kumar demonstrate that Meta RL can help LLMs better optimize test time compute

Amrith Setlur (@setlur_amrith) 's Twitter Profile Photo

How to effectively unlearn finetuning data? ❌ Approx. methods leak sensitive data ✅ Exact unlearning (eg. retraining) is secure 🔒 but inefficient 🚨 New paper: *efficient* & *exact* unlearning (led by Kevin) 🗝️ Idea: model merging at scale arxiv.org/pdf/2504.04626 🧵⤵️

How to effectively unlearn finetuning data?
❌ Approx. methods leak sensitive data
✅ Exact unlearning (eg. retraining) is secure 🔒 but inefficient

🚨 New paper: *efficient* & *exact* unlearning (led by Kevin)
🗝️ Idea: model merging at scale
arxiv.org/pdf/2504.04626
🧵⤵️
Andrew Gordon Wilson (@andrewgwils) 's Twitter Profile Photo

The ICML 2025 workshops list is online! icml.cc/virtual/2025/e…. Many exciting topics, spanning multi-agent systems, world models, test-time adaptation, actionable interpretability, and much more.

ML@CMU (@mlcmublog) 's Twitter Profile Photo

blog.ml.cmu.edu/2025/04/18/llm… 📈⚠️ Is your LLM unlearning benchmark measuring what you think it is? In a new blog post authored by Pratiksha Thaker, Shengyuan Hu, Neil Kale, Yash Maurya, Steven Wu, and Virginia Smith, we discuss why empirical benchmarks are necessary but not

Aashiq Muhamed (@aashiqmuhamed) 's Twitter Profile Photo

Thrilled to share our new work on improving LLM unlearning! 🚀 Gradient-based unlearning struggle with high cost, instability & lack of precision. We introduce Dynamic SAE Guardrails (DSG): an activation-based approach using SAEs for targeted, efficient knowledge removal.

Thrilled to share our new work on improving LLM unlearning! 🚀 
Gradient-based unlearning struggle with high cost, instability & lack of precision.
We introduce Dynamic SAE Guardrails (DSG): an activation-based approach using SAEs for targeted, efficient knowledge removal.
ICML Conference (@icmlconf) 's Twitter Profile Photo

Invited talked are announced. icml.cc/virtual/2025/e… Jon Kleinberg Pamela Samuelson Frauke Kreuter Anca Dragan Andreas Krause

ML@CMU (@mlcmublog) 's Twitter Profile Photo

blog.ml.cmu.edu/2025/05/22/unl… Are your LLMs truly forgetting unwanted data?  In this new blog post authored by Shengyuan Hu, Yiwei Fu, Steven Wu, and Virginia Smith, we discuss how benign relearning can jog unlearned LLM's memory to recover knowledge that is supposed to be forgotten.

Ameet Talwalkar (@atalwalkar) 's Twitter Profile Photo

I’m excited to share new work from Datadog AI Research! We just released Toto, a new SOTA (by a wide margin!) time series foundation model, and BOOM, the largest benchmark of observability metrics. Both are available under the Apache 2.0 license. 🧵

I’m excited to share new work from Datadog AI Research! We just released Toto, a new SOTA (by a wide margin!) time series foundation model, and BOOM, the largest benchmark of observability metrics. Both are available under the Apache 2.0 license. 🧵
Matthew Yang (@matthewyryang) 's Twitter Profile Photo

🚨 NEW PAPER: What if LLMs could tackle harder problems - not by explicitly training on longer traces, but by learning how to think longer? Our recipe e3 teaches models to explore in-context, enabling LLMs to unlock longer reasoning chains without ever seeing them in training.

ICML Conference (@icmlconf) 's Twitter Profile Photo

ICML offers an optional poster printing service icml.myprintdesk.net Orders can be picked up the day at the Vancouver Convention Centre in West MR 104 during the following hours: Monday - Friday: 7:30 am - 5:00 pm Saturday: 8:00 am - 1:00 pm

Pratiksha Thaker (@prthaker_) 's Twitter Profile Photo

I'm very excited to share some new work arxiv.org/abs/2506.06488. This work started out in conversations with Thorn where we realized that shadow model MIAs couldn't be used to audit models for harmful content of children. See 🧵 for why, and our progress on solving this...