Tim Dettmers (@tim_dettmers) 's Twitter Profile
Tim Dettmers

@tim_dettmers

Creator of bitsandbytes.Research Scientist @allen_ai and incoming professor @CarnegieMellon. I blog about deep learning and PhD life at timdettmers.com.

ID: 872274950

linkhttp://timdettmers.com/about calendar_today10-10-2012 18:18:30

3,3K Tweet

36,36K Followers

964 Following

Rulin Shao (@rulinshao) 's Twitter Profile Photo

🎉Our Spurious Rewards is available on ArXiv! We added experiments on - More prompts/steps/models/analysis... - Spurious Prompts! Surprisingly, we obtained 19.4% gains when replacing prompts with LaTex placeholder text (\lipsum) 😶‍🌫️ Check out our 2nd blog: tinyurl.com/spurious-prompt

🎉Our Spurious Rewards is available on ArXiv! We added experiments on
- More prompts/steps/models/analysis...
- Spurious Prompts!
Surprisingly, we obtained 19.4% gains when replacing prompts with LaTex placeholder text (\lipsum) 😶‍🌫️

Check out our 2nd blog: tinyurl.com/spurious-prompt
Stella Li (@stellalisy) 's Twitter Profile Photo

Spurious Rewards was not all‼️We now present spurious PROMPTS🤔 check out our latest findings and discussion on evaluation: tinyurl.com/spurious-prompt. Who knew Lorem ipsum can bring 19.4% gains compared to default prompt👀 Also, arXiv is out🤩 arxiv.org/abs/2506.10947📄

Spurious Rewards was not all‼️We now present spurious PROMPTS🤔 check out our latest findings and discussion on evaluation: tinyurl.com/spurious-prompt.

Who knew Lorem ipsum can bring 19.4% gains compared to default prompt👀

Also, arXiv is out🤩 arxiv.org/abs/2506.10947📄
Sawyer Merritt (@sawyermerritt) 's Twitter Profile Photo

Waymo in a new blog post: "We conducted a comprehensive study using Waymo’s internal dataset. Spanning 500,000 hours of driving, it is significantly larger than any dataset used in previous scaling studies in the AV domain. Our study uncovered the following: • Similar to LLMs,

Waymo in a new blog post: "We conducted a comprehensive study using Waymo’s internal dataset. Spanning 500,000 hours of driving, it is significantly larger than any dataset used in previous scaling studies in the AV domain.

Our study uncovered the following: 
• Similar to LLMs,
zhyncs (@zhyncs42) 's Twitter Profile Photo

SGLang is an early user of FlashInfer and witnessed its rise as the de facto LLM inference kernel library. It won best paper at MLSys 2025, and Zihao now leads its development NVIDIA AI Developer. SGLang’s GB200 NVL72 optimizations were made possible with strong support from the

Alex Zhang (@a1zhang) 's Twitter Profile Photo

btw a shit ton of amazing learning material + open-source code for GPU programming ($150K worth) is linked on the latest GPU MODE news post a year ago when I was an undergrad I was scouring the internet for these kinds of resources, plz take advantage of it!

btw a shit ton of amazing learning material + open-source code for GPU programming ($150K worth) is linked on the latest <a href="/GPU_MODE/">GPU MODE</a> news post

a year ago when I was an undergrad I was scouring the internet for these kinds of resources, plz take advantage of it!
Tianqi Chen (@tqchenml) 's Twitter Profile Photo

Check out our work on parallel reasoning 🧠; We bring an AI-assisted curator that identifies parallel paths in sequential traces, then tune models into native parallel thinkers that runs efficiently with prefix sharing and batching. Really excited about this general direction

Jiacheng Liu (@liujc1998) 's Twitter Profile Photo

New paper: Infini-gram now has a “mini” version! Very compressed index, 14x less storage req, we optimized for massive indexing & efficient serving. Free to use via our Web Interface and API. Has helped us unveil eval contamination at scale. Check out more in thread below ⬇️

Joel Jang (@jang_yoel) 's Twitter Profile Photo

🚀 GR00T Dreams code is live! NVIDIA GEAR Lab's open-source solution for robotics data via video world models. Fine-tune on any robot, generate 'dreams', extract actions with IDM, and train visuomotor policies with LeRobot datasets (GR00T N1.5, SmolVLA). github.com/NVIDIA/GR00T-D…

Chris Donahue (@chrisdonahuey) 's Twitter Profile Photo

Excited to announce 🎵Magenta RealTime, the first open weights music generation model capable of real-time audio generation with real-time control. 👋 **Try Magenta RT on Colab TPUs**: colab.research.google.com/github/magenta… 👀 Blog post: g.co/magenta/rt 🧵 below

Tim Dettmers (@tim_dettmers) 's Twitter Profile Photo

This is a very important initiative. With funding cuts and the David vs Goliath battle of academia vs industry, these kinds of efforts will be critical to ensure that AI can benefit people broadly across societies.

Thao Nguyen (@thao_nguyen26) 's Twitter Profile Photo

Web data, the “fossil fuel of AI”, is being exhausted. What’s next?🤔 We propose Recycling the Web to break the data wall of pretraining via grounded synthetic data. It is more effective than standard data filtering methods, even with multi-epoch repeats! arxiv.org/abs/2506.04689

Web data, the “fossil fuel of AI”, is being exhausted. What’s next?🤔
We propose Recycling the Web to break the data wall of pretraining via grounded synthetic data. It is more effective than standard data filtering methods, even with multi-epoch repeats!

arxiv.org/abs/2506.04689
Stas Bekman (@stasbekman) 's Twitter Profile Photo

My first project at Snowflake AI Research is complete! I present to you Arctic Long Sequence Training (ALST) Paper: arxiv.org/abs/2506.13996 Blog: snowflake.com/en/engineering… ALST is a set of modular, open-source techniques that enable training on sequences up to 15 million

My first project at <a href="/Snowflake/">Snowflake</a> AI Research is complete! 

I present to you Arctic Long Sequence Training (ALST) 

Paper: arxiv.org/abs/2506.13996
Blog: snowflake.com/en/engineering…

ALST is a set of modular, open-source techniques that enable training on sequences up to 15 million
Niloofar (on faculty job market!) (@niloofar_mire) 's Twitter Profile Photo

🪄We made a 1B Llama BEAT GPT-4o by... making it MORE private?! LoCoMo results: 🔓GPT-4o: 80.6% 🔐1B Llama + GPT-4o (privacy): 87.7% (+7.1!⏫) 💡How? GPT-4o provides reasoning ("If X then Y"), the local model fills in the blanks with your private data to get the answer!

🪄We made a 1B Llama BEAT GPT-4o by... making it MORE private?!

LoCoMo results:
🔓GPT-4o: 80.6% 
🔐1B Llama + GPT-4o (privacy): 87.7% (+7.1!⏫)

💡How? GPT-4o provides reasoning ("If X then Y"), the local model fills in the blanks with your private data to get the answer!
Tianqi Chen (@tqchenml) 's Twitter Profile Photo

#MLSys2026 will be led by the general chair Luis Ceze and PC chairs Zhihao Jia and Aakanksha Chowdhery. The conference will be held in Bellevue on Seattle's east side. Consider submitting and bringing your latest works in AI and systems—more details at mlsys.org.

Aditya Kusupati (@adityakusupati) 's Twitter Profile Photo

📢Now open, Gemma 3n weights & it is natively flexible, first of its kind, thanks to MatFormer🪆 Any model between E4B & E2B with ZERO training near Pareto -- we found a bunch! Find a better E3B than what we released, I will send you a 🪆😉 Find the colab for extraction 🧵👇🪆

📢Now open, Gemma 3n weights &amp; it is natively flexible, first of its kind, thanks to MatFormer🪆

Any model between E4B &amp; E2B with ZERO training near Pareto -- we found a bunch!

Find a better E3B than what we released, I will send you a 🪆😉

Find the colab for extraction 🧵👇🪆
Graham Neubig (@gneubig) 's Twitter Profile Photo

What will software development look like in 2026? With coding agents rapidly improving, dev roles may look quite different. My current workflow has changed a lot: - Work in github, not IDEs - Agents in parallel - Write English, not code - More code review Thoughts + a video👇

Julian Michael (@_julianmichael_) 's Twitter Profile Photo

I should probably announce that a few months ago, I joined Scale AI to lead the Safety, Evaluations, and Alignment Lab… and today, I joined Meta to continue working on AI alignment with Summer Yue and Alexandr Wang. Very excited for what we can accomplish together!

Albert Gu (@_albertgu) 's Twitter Profile Photo

I really like this result: an elegant framing and solution to significantly improve length generalization in recurrent models at large (RNNs/SSMs/linear attention/etc). This has significant implications for the problems architecture researchers should focus on, IMO

Rulin Shao (@rulinshao) 's Twitter Profile Photo

🚀 Last year: MassiveDS-1.4T showed great scaling gains with a web-scale datastore but was too heavy for online production ✨ Now: CompactDS is here! Better performance, compact size, ready for agentic apps & Deep Research RL training Kudos to Xinxi Lyu Michael Duan for leading this!