Belinda Li @ ICLR 2025 (@belindazli) 's Twitter Profile
Belinda Li @ ICLR 2025

@belindazli

PhD student @MIT_CSAIL | formerly SWE @facebookai, BS'19 @uwcse | NLP, ML

ID: 1188224435364327424

linkhttp://belindal.github.io calendar_today26-10-2019 22:42:43

434 Tweet

2,2K Followers

650 Following

Laura Ruis (@lauraruis) 's Twitter Profile Photo

Excited to announce that this fall I'll be joining Jacob Andreas's amazing lab at MIT for a postdoc to work on interp. for reasoning (with Ev (like in 'evidence', not Eve) Fedorenko 🇺🇦 🤯 among others). Cannot wait to think more about this direction in such a dream academic context!

Stanford NLP Group (@stanfordnlp) 's Twitter Profile Photo

For this week’s NLP Seminar, we are thrilled to host Jacob Andreas to talk about “Just Asking Questions” When: 5/15 Thurs 11am PT Non-Stanford affiliates registration form: forms.gle/svy5q5uu7anHw7…

For this week’s NLP Seminar, we are thrilled to host <a href="/jacobandreas/">Jacob Andreas</a> to talk about “Just Asking Questions”
When: 5/15 Thurs 11am PT
Non-Stanford affiliates registration form: forms.gle/svy5q5uu7anHw7…
Tiago Pimentel (@tpimentelms) 's Twitter Profile Photo

If you're finishing your camera-ready for ACL (#acl2025nlp) or ICML (#icml2025 ) and want to cite co-first authors more fairly, I just made a simple fix to do this! Just add $^*$ to the authors' names in your bibtex, and the citations should change :) github.com/tpimentelms/ac…

If you're finishing your camera-ready for ACL (#acl2025nlp) or ICML (#icml2025 ) and want to cite co-first authors more fairly, I just made a simple fix to do this! Just add $^*$ to the authors' names in your bibtex, and the citations should change :)

github.com/tpimentelms/ac…
Jon Richens (@jonathanrichens) 's Twitter Profile Photo

Are world models necessary to achieve human-level agents, or is there a model-free short-cut? Our new #ICML2025 paper tackles this question from first principles, and finds a surprising answer, agents _are_ world models… 🧵

Are world models necessary to achieve human-level agents, or is there a model-free short-cut?
Our new #ICML2025 paper tackles this question from first principles, and finds a surprising answer, agents _are_ world models… 🧵
Yijia Shao (@echoshao8899) 's Twitter Profile Photo

🚨 70 million US workers are about to face their biggest workplace transmission due to AI agents. But nobody asks them what they want. While AI races to automate everything, we took a different approach: auditing what workers want vs. what AI can do across the US workforce.🧵

🚨 70 million US workers are about to face their biggest workplace transmission due to AI agents. But nobody asks them what they want.

While AI races to automate everything, we took a different approach: auditing what workers want vs. what AI can do across the US workforce.🧵
Erik Brynjolfsson (@erikbryn) 's Twitter Profile Photo

Some tasks are painful to do. But some are fulfilling and fun. How do they line up with the tasks that AI agents are set to automate? Not that well, based on our new paper "Future of Work with AI Agents: Auditing Automation and Augmentation Potential across the U.S. Workforce"

Some tasks are painful to do.
But some are fulfilling and fun.

How do they line up with the tasks that AI agents are set to automate?

Not that well, based on our new paper "Future of Work with AI Agents: Auditing Automation and Augmentation Potential across the U.S. Workforce"
Morris Yau (@morrisyau) 's Twitter Profile Photo

Transformers: ⚡️fast to train (compute-bound), 🐌slow to decode (memory-bound). Can Transformers be optimal in both? Yes! By exploiting sequential-parallel duality. We introduce Transformer-PSM with constant time per token decode. 🧐 arxiv.org/pdf/2506.10918

Transformers: ⚡️fast to train (compute-bound), 🐌slow to decode (memory-bound).
Can Transformers be optimal in both? Yes! By exploiting sequential-parallel duality.  We introduce Transformer-PSM with constant time per token decode.  🧐   arxiv.org/pdf/2506.10918
Ilia Sucholutsky (@sucholutsky) 's Twitter Profile Photo

Thrilled to announce I’ll be joining Purdue Computer Science as an Assistant Professor in Fall 2026! My lab will work on AI thought partners, machines that think with people rather than instead of people – I'll be recruiting PhD students this upcoming cycle so reach out & apply if interested!

Thrilled to announce I’ll be joining <a href="/PurdueCS/">Purdue Computer Science</a> as an Assistant Professor in Fall 2026!

My lab will work on AI thought partners, machines that think with people rather than instead of people – I'll be recruiting PhD students this upcoming cycle so reach out &amp; apply if interested!
Laura Ruis (@lauraruis) 's Twitter Profile Photo

LLMs can be programmed by backprop 🔎 In our new preprint, we show they can act as fuzzy program interpreters and databases. After being ‘programmed’ with next-token prediction, they can retrieve, evaluate, and even *compose* programs at test time, without seeing I/O examples.

LLMs can be programmed by backprop 🔎

In our new preprint, we show they can act as fuzzy program interpreters and databases. After being ‘programmed’ with next-token prediction, they can retrieve, evaluate, and even *compose* programs at test time, without seeing I/O examples.
CLS (@chengleisi) 's Twitter Profile Photo

Are AI scientists already better than human researchers? We recruited 43 PhD students to spend 3 months executing research ideas proposed by an LLM agent vs human experts. Main finding: LLM ideas result in worse projects than human ideas.

Are AI scientists already better than human researchers?

We recruited 43 PhD students to spend 3 months executing research ideas proposed by an LLM agent vs human experts.

Main finding: LLM ideas result in worse projects than human ideas.
jessica dai (@jessicadai_) 's Twitter Profile Photo

individual reporting for post-deployment evals — a little manifesto (& new preprints!) tldr: end users have unique insights about how deployed systems are failing; we should figure out how to translate their experiences into formal evaluations of those systems.

individual reporting for post-deployment evals — a little manifesto (&amp; new preprints!)

tldr: end users have unique insights about how deployed systems are failing; we should figure out how to translate their experiences into formal evaluations of those systems.
Keyon Vafa (@keyonv) 's Twitter Profile Photo

Can an AI model predict perfectly and still have a terrible world model? What would that even mean? Our new ICML paper formalizes these questions One result tells the story: A transformer trained on 10M solar systems nails planetary orbits. But it botches gravitational laws 🧵

xuan (ɕɥɛn / sh-yen) (@xuanalogue) 's Twitter Profile Photo

Ever since I started thinking seriously about AI value alignment in 2016-7, I've been frustrated by the inadequacy of utility+RL theory to account for the richness of human values. Glad to be part of a larger team now moving beyond those thin theories towards thicker ones.

Ever since I started thinking seriously about AI value alignment in 2016-7, I've been frustrated by the inadequacy of utility+RL theory to account for the richness of human values.

Glad to be part of a larger team now moving beyond those thin theories towards thicker ones.
Belinda Li @ ICLR 2025 (@belindazli) 's Twitter Profile Photo

I'll be presenting "(How) Do Language Models Track State" at ICML! Come by our poster tomorrow, Tuesday July 15 from 4:30pm - 7pm to chat about LMs and whether/how they encode dynamic world models! 🔗 icml.cc/virtual/2025/p…

I'll be presenting "(How) Do Language Models Track State" at ICML!
Come by our poster tomorrow, Tuesday July 15 from 4:30pm - 7pm to chat about LMs and whether/how they encode dynamic world models!

🔗 icml.cc/virtual/2025/p…
ICML 2025 Workshop on Assessing World Models (@worldmodelsicml) 's Twitter Profile Photo

Join us for the Workshop on Assessing World Models at ICML tomorrow! When: Friday July 17, 8:45am-5:15pm Where: West Ballroom B (same floor as registration)

Join us for the Workshop on Assessing World Models at ICML tomorrow!

When: Friday July 17, 8:45am-5:15pm
Where: West Ballroom B (same floor as registration)
Belinda Li @ ICLR 2025 (@belindazli) 's Twitter Profile Photo

Come check out our "Assessing World Models" workshop tomorrow! We'll be discussing whether generative AI builds world models, and what these world models might look like.

Tyler Brooke-Wilson (@t_brookewilson) 's Twitter Profile Photo

How do people reason while still staying coherent – as if they have an internal ‘world model’ for situations they’ve never encountered? A new paper on open-world cognition (preview at the world models workshop at #ICML2025!)

How do people reason while still staying coherent – as if they have an internal ‘world model’ for situations they’ve never encountered? A new paper on open-world cognition (preview at the world models workshop at #ICML2025!)
Jiaxin Pei (@jiaxin_pei) 's Twitter Profile Photo

Life Update: I will join School of Information - UT Austin as an Assistant Professor in Fall 2026 and will continue my work on LLM, HCI, and Computational Social Science. I'm building a new lab on Human-Centered AI Systems and will be hiring PhD students in the coming cycle!

Life Update: I will join <a href="/UTiSchool/">School of Information - UT Austin</a> as an Assistant Professor in Fall 2026 and will continue my work on LLM, HCI, and Computational Social Science. I'm building a new lab on Human-Centered AI Systems and will be hiring PhD students in the coming cycle!
Katie Collins (@katie_m_collins) 's Twitter Profile Photo

How do people reason so flexibly about new problems, bringing to bear globally-relevant knowledge while staying locally-consistent? Can we engineer a system that can synthesize bespoke world models (expressed as probabilistic programs) on-the-fly?

How do people reason so flexibly about new problems, bringing to bear globally-relevant knowledge while staying locally-consistent? Can we engineer a system that can synthesize bespoke world models (expressed as probabilistic programs) on-the-fly?
Mehul Damani @ ICLR (@mehuldamani2) 's Twitter Profile Photo

🚨New Paper!🚨 We trained reasoning LLMs to reason about what they don't know. o1-style reasoning training improves accuracy but produces overconfident models that hallucinate more. Meet RLCR: a simple RL method that trains LLMs to reason and reflect on their uncertainty --

🚨New Paper!🚨
We trained reasoning LLMs to reason about what they don't know.

o1-style reasoning training improves accuracy but produces overconfident models that hallucinate more.

Meet RLCR: a simple RL method that trains LLMs to reason and reflect on their uncertainty --