Sewon Min (@sewon__min) 's Twitter Profile
Sewon Min

@sewon__min

Incoming faculty @Berkeley_EECS @berkeley_ai || Research scientist at @allen_ai || PhD from @uwcse @uwnlp

ID: 928289778960707584

linkhttp://sewonmin.com calendar_today08-11-2017 15:55:04

972 Tweet

11,11K Followers

739 Following

Berkeley AI Research (@berkeley_ai) 's Twitter Profile Photo

BAIR faculty Stuart Russell, Dan Klein, Alane Suhr, Ken Goldberg, and Sewon Min weigh in on the future of LLMs, synthetic data, and the road ahead ⬇️ alumni.berkeley.edu/california-mag…

Rulin Shao (@rulinshao) 's Twitter Profile Photo

Meet ReasonIR-8B✨the first retriever specifically trained for reasoning tasks! Our challenging synthetic training data unlocks SOTA scores on reasoning IR and RAG benchmarks. ReasonIR-8B ranks 1st on BRIGHT and outperforms search engine and retriever baselines on MMLU and GPQA🔥

Meet ReasonIR-8B✨the first retriever specifically trained for reasoning tasks! Our challenging synthetic training data unlocks SOTA scores on reasoning IR and RAG benchmarks. ReasonIR-8B ranks 1st on BRIGHT and outperforms search engine and retriever baselines on MMLU and GPQA🔥
Ai2 (@allen_ai) 's Twitter Profile Photo

We’re live on Reddit! Ask us Anything about our OLMo family of models. We have six of our researchers on hand to answer all your questions.

We’re live on Reddit! Ask us Anything about our OLMo family of models. We have six of our researchers on hand to answer all your questions.
Muhammad Khalifa (@mkhalifaaaa) 's Twitter Profile Photo

🚨Announcing SCALR @ COLM 2025 — Call for Papers!🚨 The 1st Workshop on Test-Time Scaling and Reasoning Models (SCALR) is coming to Conference on Language Modeling in Montreal this October! This is the first workshop dedicated to this growing research area. 🌐 scalr-workshop.github.io

🚨Announcing SCALR @ COLM 2025 — Call for Papers!🚨

The 1st Workshop on Test-Time Scaling and Reasoning Models (SCALR) is coming to <a href="/COLM_conf/">Conference on Language Modeling</a>  in Montreal this October!

This is the first workshop dedicated to this growing research area.

🌐 scalr-workshop.github.io
Stella Li (@stellalisy) 's Twitter Profile Photo

🤯 We cracked RLVR with... Random Rewards?! Training Qwen2.5-Math-7B with our Spurious Rewards improved MATH-500 by: - Random rewards: +21% - Incorrect rewards: +25% - (FYI) Ground-truth rewards: + 28.8% How could this even work⁉️ Here's why: 🧵 Blogpost: tinyurl.com/spurious-rewar…

🤯 We cracked RLVR with... Random Rewards?!
Training Qwen2.5-Math-7B with our Spurious Rewards improved MATH-500 by:
- Random rewards: +21%
- Incorrect rewards: +25%
- (FYI) Ground-truth rewards: + 28.8%
How could this even work⁉️ Here's why: 🧵
Blogpost: tinyurl.com/spurious-rewar…
Omar Khattab (@lateinteraction) 's Twitter Profile Photo

Tensor Templar The important breakthrough is that a lot of the “RL just works” noise has little to do with RL and has more to do with “qwen mid-training for math and coding makes the model very receptive for the same jumps on math and coding over and over”

Yizhong Wang (@yizhongwyz) 's Twitter Profile Photo

Thrilled to announce that I will be joining UT Austin Computer Science at UT Austin as an assistant professor in fall 2026! I will continue working on language models, data challenges, learning paradigms, & AI for innovation. Looking forward to teaming up with new students & colleagues! 🤠🤘

Thrilled to announce that I will be joining <a href="/UTAustin/">UT Austin</a> <a href="/UTCompSci/">Computer Science at UT Austin</a> as an assistant professor in fall 2026! 

I will continue working on language models, data challenges, learning paradigms, &amp; AI for innovation. Looking forward to teaming up with new students &amp; colleagues! 🤠🤘
Allen School (@uwcse) 's Twitter Profile Photo

Congratulations to University of Washington #UWAllen Ph.D. grads Ashish Sharma & Sewon Min, Association for Computing Machinery Doctoral Dissertation Award honorees! Sharma won for #AI tools for mental health; Min received honorable mention for efficient, flexible language models. #ThisIsUW news.cs.washington.edu/2025/06/04/all…

Association for Computing Machinery (@theofficialacm) 's Twitter Profile Photo

🎓 Congrats to Ashish Sharma, University of Washington on receiving the ACM Doctoral Dissertation Award for his dissertation, "Human-AI Collaboration to Support Mental Health and Well Being." 👏 Honorable Mentions: Alexander Kelley, University of Illinois Sewon Min, UC Berkeley

🎓 Congrats to Ashish Sharma, <a href="/UW/">University of Washington</a> on receiving the ACM Doctoral Dissertation Award for his dissertation, "Human-AI Collaboration to Support Mental Health and Well Being." 

👏 Honorable Mentions:
Alexander Kelley, <a href="/UofIllinois/">University of Illinois</a>
Sewon Min, <a href="/UCBerkeley/">UC Berkeley</a>
EleutherAI (@aieleuther) 's Twitter Profile Photo

Can you train a performant language models without using unlicensed text? We are thrilled to announce the Common Pile v0.1, an 8TB dataset of openly licensed and public domain text. We train 7B models for 1T and 2T tokens and match the performance similar models like LLaMA 1&2

Can you train a performant language models without using unlicensed text?

We are thrilled to announce the Common Pile v0.1, an 8TB dataset of openly licensed and public domain text. We train 7B models for 1T and 2T tokens and match the performance similar models like LLaMA 1&amp;2
Sergey Levine (@svlevine) 's Twitter Profile Photo

I always found it puzzling how language models learn so much from next-token prediction, while video models learn so little from next frame prediction. Maybe it's because LLMs are actually brain scanners in disguise. Idle musings in my new blog post: sergeylevine.substack.com/p/language-mod…

Ary (@aryg18) 's Twitter Profile Photo

fwiw, I think Prof. Percy Liang and the CS336 team nailed this: Sutton’s Bitter Lesson is often misinterpreted as “scale is all that matters” and/or “algorithms don’t matter.” The more accurate – and useful – interpretation is: what matters are the algorithms that scale.

Rulin Shao (@rulinshao) 's Twitter Profile Photo

🚀 Last year: MassiveDS-1.4T showed great scaling gains with a web-scale datastore but was too heavy for online production ✨ Now: CompactDS is here! Better performance, compact size, ready for agentic apps & Deep Research RL training Kudos to Xinxi Lyu Michael Duan for leading this!

Weijia Shi (@weijiashi2) 's Twitter Profile Photo

Can data owners & LM developers collaborate to build a strong shared model while each retaining data control? Introducing FlexOlmo💪, a mixture-of-experts LM enabling: • Flexible training on your local data without sharing it • Flexible inference to opt in/out your data

Akari Asai (@akariasai) 's Twitter Profile Photo

Some updates 🚨 I finished my Ph.D at Allen School in June 2025! After a year at AI2 as a Research Scientist, I am joining CMU Language Technologies Institute | @CarnegieMellon & Machine Learning Dept. at Carnegie Mellon (courtesy) as an Assistant Professor in Fall 2026. The journey, acknowledgments & recruiting in 🧵

Some updates 🚨
I finished my Ph.D at <a href="/uwcse/">Allen School</a> in June 2025!
After a year at AI2 as a Research Scientist, I am joining CMU <a href="/LTIatCMU/">Language Technologies Institute | @CarnegieMellon</a> &amp; <a href="/mldcmu/">Machine Learning Dept. at Carnegie Mellon</a> (courtesy) as an Assistant Professor in Fall 2026.
The journey, acknowledgments &amp; recruiting in 🧵