Yiqing Xie (@yiqingxienlp) 's Twitter Profile
Yiqing Xie

@yiqingxienlp

โœจ Annotation-efficient Train/Eval; Code-Gen;

๐ŸŽ“ PhD student @LTIatCMU; MSCS @dmguiuc.
๐Ÿ‘ฉโ€๐Ÿ’ป previously Intern @meta; @MSFTResearch; @AlibabaDAMO.

ID: 1698061163966210049

linkhttps://yiqingxyq.github.io/ calendar_today02-09-2023 19:52:05

60 Tweet

157 Followers

183 Following

Pranjal Aggarwal (@pranjalaggarw16) 's Twitter Profile Photo

What if you could control how long a reasoning model โ€œthinksโ€? Presenting L1-1.5B, an RL-trained reasoning model with: - controllable thinking length via a prompt - better performance per token than S1 - better short CoT performance than GPT-4o cmu-l3.github.io/l1 ๐Ÿงต

Jacob Springer (@jacspringer) 's Twitter Profile Photo

Training with more data = better LLMs, right? ๐Ÿšจ False! Scaling language models by adding more pre-training data can decrease your performance after post-training! Introducing "catastrophic overtraining." ๐Ÿฅ๐Ÿงต+arXiv ๐Ÿ‘‡ 1/9

Training with more data = better LLMs, right? ๐Ÿšจ

False! Scaling language models by adding more pre-training data can decrease your performance after post-training!

Introducing "catastrophic overtraining." ๐Ÿฅ๐Ÿงต+arXiv ๐Ÿ‘‡

1/9
Gashon Hussein (@gashonhussein) 's Twitter Profile Photo

Excited to share our new paper, "One-Minute Video Generation with Test-Time Training (TTT)" in collaboration with NVIDIA. We augment a pre-trained Transformer with TTT-layers and finetune it to generate one-minute Tom and Jerry cartoons with strong temporal and spatial

Excited to share our new paper, "One-Minute Video Generation with Test-Time Training (TTT)" in collaboration with NVIDIA.

We augment a pre-trained Transformer with TTT-layers and finetune it to generate one-minute Tom and Jerry cartoons with strong temporal and spatial
Shubham Gandhi (@shubhamrgandhi) 's Twitter Profile Photo

๐ŸšจNew preprint๐Ÿšจ Iโ€™m super excited to share our work: An Empirical Study on Strong-Weak Model Collaboration for Repo-level Code Generation ๐Ÿ“œ: arxiv.org/abs/2505.20182 w/ Atharva Naik , Yiqing Xie and Carolyn Rose ๐Ÿงต

Leena Mathur (@lmathur_) 's Twitter Profile Photo

Future AI systems interacting with humans will need to perform social reasoning that is grounded in behavioral cues and external knowledge. We introduce Social Genome to study and advance this form of reasoning in models! New paper w/ Marian Qian, Paul Liang, & LP Morency!

Future AI systems interacting with humans will need to perform social reasoning that is grounded in behavioral cues and external knowledge. 

We introduce Social Genome to study and advance this form of reasoning in models!

New paper w/ Marian Qian, <a href="/pliang279/">Paul Liang</a>, &amp; <a href="/lpmorency/">LP Morency</a>!
Xuhui Zhou (@nlpxuhui) 's Twitter Profile Photo

Very excited to share that HAICosystem has been accepted to #COLM2025 ! ๐ŸŽ‰ Multi-turn, interactive evaluation is THE future, think Tau-Bench, TheAgentCompany, Sotopia, ... Proud to take a small step toward open-ended, interactive AI safety eval, and excited for whatโ€™s next! ๐Ÿ˜Ž

Daniel Fried (@dan_fried) 's Twitter Profile Photo

At this morning's #colm2025 poster session, come see Yiqing Xie present a method for scalable construction of coding environments from GitHub repos. Poster 76 Paper: arxiv.org/abs/2503.07358 Work with Alex Xie, Divyanshu Sheth, Pengfei Liu, and Carolyn Rose

Yueqi Song (@yueqi_song) 's Twitter Profile Photo

We just built and released the largest dataset for supervised fine-tuning of agentic LMs, 1.27M trajectories (~36B tokens)! Up until now, large-scale SFT for agents is rare - not for lack of data, but because of fragmentation across heterogeneous formats, tools, and interfaces.