Brando Miranda (@brandohablando) 's Twitter Profile
Brando Miranda

@brandohablando

CS Ph.D. @Stanford, researching data quality, foundation models, and ML for Theorem Proving. Prev: @MIT, @MIT_CBMM, @IllinoisCS, @IBM. Opinions are mine. 🇲🇽

ID: 1253358235

linkhttps://brando90.github.io/brandomiranda/publications.html calendar_today09-03-2013 04:01:37

1,1K Tweet

963 Followers

736 Following

Percy Liang (@percyliang) 's Twitter Profile Photo

Wrapped up Stanford CS336 (Language Models from Scratch), taught with an amazing team Tatsunori Hashimoto Marcel Rød Neil Band Rohith Kuditipudi. Researchers are becoming detached from the technical details of how LMs work. In CS336, we try to fix that by having students build everything:

Rylan Schaeffer (@rylanschaeffer) 's Twitter Profile Photo

Another #ICML2025 paper! Why Has Predicting Downstream Capabilities of Frontier AI Models with Scale Remained Elusive? TLDR: Predicting language model performance with scale on multiple choice question-answer (MCQA) benchmarks is made difficult b/c ... 1/3

Another #ICML2025 paper!

Why Has Predicting Downstream Capabilities of Frontier AI Models with Scale Remained Elusive?

TLDR: Predicting language model performance with scale on multiple choice question-answer (MCQA) benchmarks is made difficult b/c ...

1/3
Rylan Schaeffer (@rylanschaeffer) 's Twitter Profile Photo

Third #ICML2025 paper! What effect will web-scale synthetic data have on future deep generative models? Collapse or Thrive? Perils and Promises of Synthetic Data in a Self-Generating World 🔄 Joshua Kazdan Apratim Dey Matthias Gerstgrasser Rafael Rafailov @ NeurIPS Sanmi Koyejo 1/7

Third #ICML2025 paper! What effect will web-scale synthetic data have on future deep generative models?

Collapse or Thrive? Perils and Promises of Synthetic Data in a Self-Generating World 🔄

<a href="/JoshuaK92829/">Joshua Kazdan</a> <a href="/ApratimDey2/">Apratim Dey</a> <a href="/MGerstgrasser/">Matthias Gerstgrasser</a> <a href="/rm_rafailov/">Rafael Rafailov @ NeurIPS</a> <a href="/sanmikoyejo/">Sanmi Koyejo</a> 

1/7
Rylan Schaeffer (@rylanschaeffer) 's Twitter Profile Photo

Joshua Kazdan Lastly, don't sleep on our NEW position paper: Model Collapse Does Not Mean What You Think We discuss the state of research on synthetic data & model collapse, and where we feel more effort is necessary w/ Alvan Arulandu Joshua Kazdan Sanmi Koyejo arxiv.org/abs/2503.03150 7/7

<a href="/JoshuaK92829/">Joshua Kazdan</a> Lastly, don't sleep on our NEW position paper: Model Collapse Does Not Mean What You Think

We discuss the state of research on synthetic data &amp; model collapse, and where we feel more effort is necessary

w/ <a href="/AlvanArulandu/">Alvan Arulandu</a> <a href="/JoshuaK92829/">Joshua Kazdan</a> <a href="/sanmikoyejo/">Sanmi Koyejo</a> 

arxiv.org/abs/2503.03150

7/7
Lean (@leanprover) 's Twitter Profile Photo

Really enjoyed this talk by Harry Goldstein that demonstrates inventive uses of the #LeanLang InfoView enhanced by metaprogramming techniques to display real-time testing data. #LeanProver #Metaprogramming #VSCode #PropertyTesting

Really enjoyed this talk by Harry Goldstein that demonstrates inventive uses of the #LeanLang InfoView enhanced by metaprogramming techniques to display real-time testing data.

#LeanProver #Metaprogramming #VSCode #PropertyTesting
Brando Miranda (@brandohablando) 's Twitter Profile Photo

Tuğrulcan Elmas 🇹 🇯 Rylan Schaeffer Sanmi Koyejo Joshua Kazdan Yegor Denisov-Blanch Francesco Orabona Koustuv Sinha Jessica Zosa Forde Jesse Dodge Susan Zhang Matthias Gerstgrasser isha Elyas Obbad One fix: invite the peers to be authors. Being generous & correcting the field is most important. I'd be happy to write refutations for my papers. Discussion & truth matter most. The shift socially should be: being more open to corrections than feeling attacked so easily.

Rylan Schaeffer (@rylanschaeffer) 's Twitter Profile Photo

I'll be at ICML Conference #ICML2025 next week to present three papers - reach out if you want to chat about generative AI, scaling laws, synthetic data or any other AI topic! #1 How Do Large Language Monkeys Get Their Power (Laws)? x.com/RylanSchaeffer…

I'll be at <a href="/icmlconf/">ICML Conference</a> #ICML2025 next week to present three papers - reach out if you want to chat about generative AI, scaling laws, synthetic data or any other AI topic!

#1 How Do Large Language Monkeys Get Their Power (Laws)?

x.com/RylanSchaeffer…
Lilian Weng (@lilianweng) 's Twitter Profile Photo

I still find it mysterious whether and how intelligence and capabilities transfer between domains and skills - from meta learning during early days to more recent question like whether solving maths helps writing a good essay. Sometime I feel a bit pessimistic given not enough