Mehul Damani @ ICLR (@mehuldamani2) 's Twitter Profile
Mehul Damani @ ICLR

@mehuldamani2

PhD Student at MIT | Reinforcement Learning, NLP

ID: 2784283957

linkhttps://damanimehul.github.io/ calendar_today01-09-2014 15:08:58

30 Tweet

227 Followers

274 Following

Linlu Qiu (@linluqiu) 's Twitter Profile Photo

It was a great pleasure working on this project with amazing collaborators! Excited to see more opportunities opened up by scaling test-time compute!

Ekin Akyürek (@akyurekekin) 's Twitter Profile Photo

Thanks for the attention, couple important points: 1) See Jack Cole, their team is the first one who applied method privately and they get the 1st rank in the competition. 2) See the concurrent work as well: x.com/ellisk_kellis/… 3) Obviously this is not AGI, it's a

Noam Brown (@polynoamial) 's Twitter Profile Photo

With OpenAI o1, we developed one way to scale test-time compute, but it isn't the only way and might not be the best way. I'm excited to see academic researchers explore new approaches in this direction.

Seungwook Han (@seungwookh) 's Twitter Profile Photo

🧩 Why do task vectors exist in pretrained LLMs? Our new research uncovers how transformers form internal abstractions and the mechanisms behind in-context learning(ICL).

🧩  Why do task vectors exist in pretrained LLMs? 

Our new research uncovers how transformers form internal abstractions and the mechanisms behind in-context learning(ICL).
Isha Puri (@ishapuri101) 's Twitter Profile Photo

[1/x] can we scale small, open LMs to o1 level? Using classical probabilistic inference methods, YES! Joint MIT CSAIL / Red Hat AI Innovation Team work introduces a particle filtering approach to scaling inference w/o any training! check out …abilistic-inference-scaling.github.io

[1/x] can we scale small, open LMs to o1 level? Using classical probabilistic inference methods, YES! Joint <a href="/MIT_CSAIL/">MIT CSAIL</a> / <a href="/RedHat/">Red Hat</a> AI Innovation Team work introduces a particle filtering approach to scaling inference w/o any training! check out …abilistic-inference-scaling.github.io
Jeremy Bernstein (@jxbz) 's Twitter Profile Photo

I just wrote my first blog post in four years! It is called "Deriving Muon". It covers the theory that led to Muon and how, for me, Muon is a meaningful example of theory leading practice in deep learning (1/11)

I just wrote my first blog post in four years! It is called "Deriving Muon". It covers the theory that led to Muon and how, for me, Muon is a meaningful example of theory leading practice in deep learning

(1/11)
idan shenfeld (@idanshenfeld) 's Twitter Profile Photo

The next frontier for AI shouldn’t just be generally helpful. It should be helpful for you! Our new paper shows how to personalize LLMs — efficiently, scalably, and without retraining. Meet PReF (arxiv.org/abs/2503.06358) 1\n

Mehul Damani @ ICLR (@mehuldamani2) 's Twitter Profile Photo

I am super excited to be presenting our work on adaptive inference -time compute at ICLR! Come chat with me on Thursday 4/24 at 3PM (Poster #219). I am also happy to chat about RL/reasoning/ RLHF/ inference scaling (DMs are open)!

Akarsh Kumar (@akarshkumar0101) 's Twitter Profile Photo

Excited to share our position paper on the Fractured Entangled Representation (FER) Hypothesis! We hypothesize that the standard paradigm of training networks today — while producing impressive benchmark results — is still failing to create a well-organized internal

Adam Zweiger (@adamzweiger) 's Twitter Profile Photo

Come check out our ICML poster on combining Test-Time Training and In-Context Learning for on-the-fly adaptation to novel tasks like ARC-AGI puzzles. I will be presenting with Jyo Pari at E-2702, Tuesday 11-1:30!

Come check out our ICML poster on combining Test-Time Training and In-Context Learning for on-the-fly adaptation to novel tasks like ARC-AGI puzzles.

I will be presenting with <a href="/jyo_pari/">Jyo Pari</a> at E-2702, Tuesday 11-1:30!
Jacob Andreas (@jacobandreas) 's Twitter Profile Photo

👉 New preprint! Today, many the biggest challenges in LM post-training aren't just about correctness, but rather consistency & coherence across interactions. This paper tackles some of these issues by optimizing reasoning LMs for calibration rather than accuracy...