Eric Zelikman (@ericzelikman) 's Twitter Profile
Eric Zelikman

@ericzelikman

looking up @xAI // was phd-ing @stanford

ID: 137463715

linkhttp://zelikman.me calendar_today26-04-2010 20:49:42

838 Tweet

18,18K Followers

1,1K Following

Eric Zelikman (@ericzelikman) 's Twitter Profile Photo

cool pipeline for analyzing lots of screenshot data 🖼️ we need good tools to understand how we interact w/ complex algos

Stephanie Chan (@scychan_brains) 's Twitter Profile Photo

Check out our new work: Generalization from context often outperforms generalization from finetuning. And you might get the best of both worlds by spending extra compute at train-time.

Eric Zelikman (@ericzelikman) 's Twitter Profile Photo

seems like a big theme lately (e.g. also "RL for Reasoning w/ One Training Example") is that approaches don't get nearly enough bang for each training point's buck - cool!

John Yang (@jyangballin) 's Twitter Profile Photo

40% with just 1 try per task: SWE-agent-LM-32B is the new #1 open source model on SWE-bench Verified. We built it by synthesizing a ton of agentic training data from 100+ Python repos. Today we’re open-sourcing the toolkit that made it happen: SWE-smith.

40% with just 1 try per task: SWE-agent-LM-32B is the new #1 open source model on SWE-bench Verified.

We built it by synthesizing a ton of agentic training data from 100+ Python repos.

Today we’re open-sourcing the toolkit that made it happen: SWE-smith.
Eric Zelikman (@ericzelikman) 's Twitter Profile Photo

fun note: heiner once described my env config as "the final boss of python venv issues" -- has been mostly issue free for a few months now, thanks mostly to uv 🤞

noahdgoodman (@noahdgoodman) 's Twitter Profile Photo

It turns out that a lot of the most interesting behavior of LLMs can be explained without knowing anything about architecture or learning algorithms. Here we predict the rise (and fall) of in-context learning using hierarchical Bayesian methods.

Shirley Wu (@shirleyyxwu) 's Twitter Profile Photo

CollabLLM won #ICML2025 ✨Outstanding Paper Award along with 6 other works! icml.cc/virtual/2025/a… 🫂 Absolutey honored and grateful for coauthors Microsoft Research Stanford AI Lab and friends who made this happen! 🗣️ Welcome people to our presentations about CollabLLM tomorrow

Kaiyu Yang (@kaiyuyang4) 's Twitter Profile Photo

🚀 Excited to share that the Workshop on Mathematical Reasoning and AI (MATH‑AI) will be at NeurIPS 2025! 📅 Dec 6 or 7 (TBD), 2025 🌴 San Diego, California

🚀 Excited to share that the Workshop on Mathematical Reasoning and AI (MATH‑AI) will be at NeurIPS 2025!
📅 Dec 6 or 7 (TBD), 2025
🌴 San Diego, California
Shirley Wu (@shirleyyxwu) 's Twitter Profile Photo

✨ Optimas is fully open-sourced at github.com/snap-stanford/… 🏋🏻 Contribute by: 1) Submitting PRs for your own compound AI systems. 2) Extending the optimization infra. 3) Creating issues (we're happy to help optimize your systems, explore why they work or fail, and go from there

Diyi Yang (@diyi_yang) 's Twitter Profile Photo

🚨 Yanzhe Zhang's new work uses a search-based framework that evolves both sides of the game to find privacy risks --- alternating between improving attacker & defender instructions in privacy-critical LLM agent interactions!!! Attack agents evolve from blunt data grabs to

🚨 <a href="/StevenyzZhang/">Yanzhe Zhang</a>'s new work uses a search-based framework that evolves both sides of the game to find privacy risks --- alternating between improving attacker &amp; defender instructions in privacy-critical LLM agent interactions!!!

Attack agents evolve from blunt data grabs to
Andrej Karpathy (@karpathy) 's Twitter Profile Photo

I am (slowly) re-reading the Tolkien legendarium (of which Lord of the Rings is a small part). The whole body of work is so incredible and there's nothing else like it... it dilutes other worlds of fiction. Wait - your story doesn't have a comprehensive history/mythology spanning