Yiding Jiang (@yidingjiang) 's Twitter Profile
Yiding Jiang

@yidingjiang

PhD student @mldcmu @SCSatCMU. Formerly intern @MetaAI, AI resident @GoogleAI. BS from @Berkeley_EECS. Trying to understand stuff.

ID: 4515396858

linkhttp://yidingjiang.github.io calendar_today10-12-2015 07:06:19

301 Tweet

1,1K Followers

559 Following

𝚐𝔪𝟾𝚡𝚡𝟾 (@gm8xx8) 's Twitter Profile Photo

Looking beyond the next token TRELAWNEY inserts future tokens <T>...</T> during training to teach models to plan ahead—boosting reasoning, coherence, and control. Highlights: - NO ARCHITECTURE CHANGES. JUST SMARTER DATA. - works with standard decoding - enables controllable

Looking beyond the next token

TRELAWNEY inserts future tokens &lt;T&gt;...&lt;/T&gt; during training to teach models to plan ahead—boosting reasoning, coherence, and control.

Highlights:
- NO ARCHITECTURE CHANGES. JUST SMARTER DATA.
- works with standard decoding
- enables controllable
Christina Baek (@_christinabaek) 's Twitter Profile Photo

Are current reasoning models optimal for test-time scaling? 🌠 No! Models make the same incorrect guess over and over again. We show that you can fix this problem w/o any crazy tricks 💫 – just do weight ensembling (WiSE-FT) for big gains on math! 1/N

Are current reasoning models optimal for test-time scaling? 🌠
No! Models make the same incorrect guess over and over again.

We show that you can fix this problem w/o any crazy tricks 💫 – just do weight ensembling (WiSE-FT) for big gains on math!

1/N
Allan Zhou (@allanzhou17) 's Twitter Profile Photo

Excited to be presenting ADO next week at #ICLR2025! Check out a new blogpost we wrote that summarizes the key ideas and results (link below):

Sadhika Malladi (@sadhikamalladi) 's Twitter Profile Photo

Check out our online data selection alg ADO at ICLR 2025! And take a look at this blog post by Yiding Jiang and Allan Zhou summarizing the key ideas: bland.website/notes/ado/

Yutong (Kelly) He (@electronickale) 's Twitter Profile Photo

✨ Love 4o-style image generation but prefer to use Midjourney? Tired of manual prompt crafting from inspo images? PRISM to the rescue! 🖼️→📝→🖼️ We automate black-box prompt engineering—no training, no embeddings, just accurate, readable prompts from your inspo images! 1/🧵

Yiding Jiang (@yidingjiang) 's Twitter Profile Photo

Data selection and curriculum learning can be formally viewed as a compression protocol via prequential coding. New blog (with Allan Zhou ) about this neat idea that motivated ADO but didn’t make it into the paper. yidingjiang.github.io/blog/post/curr…

Allan Zhou (@allanzhou17) 's Twitter Profile Photo

How should we order training examples? In a new blogpost (w/ Yiding Jiang), we explore a compression-based perspective: order your dataset to minimize its prequential codelength.

Yiding Jiang (@yidingjiang) 's Twitter Profile Photo

A mental model I find useful: all data acquisition (web scrapes, synthetic data, RL rollouts, etc.) is really an exploration problem 🔍. This perspective has some interesting implications for where AI is heading. Wrote down some thoughts: yidingjiang.github.io/blog/post/expl…

Minqi Jiang (@minqijiang) 's Twitter Profile Photo

Recently, there has been a lot of talk of LLM agents automating ML research itself. If Llama 5 can create Llama 6, then surely the singularity is just around the corner. How can we get a pulse check on whether current LLMs are capable of driving this kind of total

Recently, there has been a lot of talk of LLM agents automating ML research itself. If Llama 5 can create Llama 6, then surely the singularity is just around the corner. 

How can we get a pulse check on whether current LLMs are capable of driving this kind of total
Jean de Nyandwi (@jeande_d) 's Twitter Profile Photo

Good blog on "era of exploration" - Data scarcity is the new bottleneck. LLMs consume data far faster than humans can produce it. We're running out of high-quality training data. - Pretraining solved exploration by accident. Pretraining effectively pays a massive, upfront

Good blog on "era of exploration"

- Data scarcity is the new bottleneck. LLMs consume data far faster than humans can produce it. We're running out of high-quality training data.

- Pretraining solved exploration by accident. Pretraining effectively pays a massive, upfront
Aya Somai (@aya_somai_) 's Twitter Profile Photo

My favorite reading of the week by Yiding Jiang: Next era is not about learning from data but deciding what data to learn from. yidingjiang.github.io/blog/post/expl…

Alex Robey (@alexrobey23) 's Twitter Profile Photo

On Monday, I'll be presenting a tutorial on jailbreaking LLMs + the security of AI agents with Hamed Hassani and Amin Karbasi at ICML. I'll be in Vancouver all week -- send me a DM if you'd like to chat about jailbreaking, AI agents, robots, distillation, or anything else!

Vaishnavh Nagarajan (@_vaishnavh) 's Twitter Profile Photo

Today Chen Wu and I will be presenting our #ICML work on creativity in the Oral 3A Reasoning session (West Exhibition Hall C) 10 - 11 am PT Or please stop by our poster right after @ East Exhibition Hall A-B #E-2505 11am-1:30pm. (Hope you enjoy some silly human drawings!)

Today <a href="/ChenHenryWu/">Chen Wu</a> and I will be presenting our #ICML work on creativity in the Oral 3A Reasoning  session (West Exhibition Hall C) 10 - 11 am PT

Or please stop by our poster right after @ East Exhibition Hall A-B #E-2505 11am-1:30pm. (Hope you enjoy some silly human drawings!)