Tu Vu (@tuvllms) 's Twitter Profile
Tu Vu

@tuvllms

Research Scientist @GoogleDeepMind & Assistant Professor @VT_CS. PhD from @UMass_NLP. Google FLAMe/FreshLLMs/Flan-T5 Collection/SPoT #NLProc

ID: 850104281197928454

linkhttps://tuvllms.github.io calendar_today06-04-2017 21:53:49

1,1K Tweet

3,3K Followers

939 Following

Mistral AI (@mistralai) 's Twitter Profile Photo

Announcing Magistral, our first reasoning model designed to excel in domain-specific, transparent, and multilingual reasoning.

Arie Cattan (@ariecattan) 's Twitter Profile Photo

🚨 RAG is a popular approach but what happens when the retrieved sources provide conflicting information?🤔 We're excited to introduce our paper: “DRAGged into CONFLICTS: Detecting and Addressing Conflicting Sources in Search-Augmented LLMs”🚀 A thread 🧵👇

🚨 RAG is a popular approach but what happens when the retrieved sources provide conflicting information?🤔

We're excited to introduce our paper: 
“DRAGged into CONFLICTS: Detecting and Addressing Conflicting Sources in Search-Augmented LLMs”🚀

A thread 🧵👇
Sakana AI (@sakanaailabs) 's Twitter Profile Photo

We’re excited to introduce Text-to-LoRA: a Hypernetwork that generates task-specific LLM adapters (LoRAs) based on a text description of the task. Catch our presentation at #ICML2025! Paper: arxiv.org/abs/2506.06105 Code: github.com/SakanaAI/Text-… Biological systems are capable of

Alex Turner (@turn_trout) 's Twitter Profile Photo

Thought real machine unlearning was impossible? We show that distilling a conventionally “unlearned” model creates a model resistant to relearning attacks. 𝐃𝐢𝐬𝐭𝐢𝐥𝐥𝐚𝐭𝐢𝐨𝐧 𝐦𝐚𝐤𝐞𝐬 𝐮𝐧𝐥𝐞𝐚𝐫𝐧𝐢𝐧𝐠 𝐫𝐞𝐚𝐥.

Thought real machine unlearning was impossible? We show that distilling a conventionally “unlearned” model creates a model resistant to relearning attacks. 𝐃𝐢𝐬𝐭𝐢𝐥𝐥𝐚𝐭𝐢𝐨𝐧 𝐦𝐚𝐤𝐞𝐬 𝐮𝐧𝐥𝐞𝐚𝐫𝐧𝐢𝐧𝐠 𝐫𝐞𝐚𝐥.
MiniMax (official) (@minimax__ai) 's Twitter Profile Photo

Day 1/5 of #MiniMaxWeek: We’re open-sourcing MiniMax-M1, our latest LLM — setting new standards in long-context reasoning. - World’s longest context window: 1M-token input, 80k-token output - State-of-the-art agentic use among open-source models - RL at unmatched efficiency:

Day 1/5 of #MiniMaxWeek: We’re open-sourcing MiniMax-M1, our latest LLM — setting new standards in long-context reasoning.

- World’s longest context window: 1M-token input, 80k-token output
- State-of-the-art agentic use among open-source models
- RL at unmatched efficiency:
Sundar Pichai (@sundarpichai) 's Twitter Profile Photo

Gemini 2.5 Pro + 2.5 Flash are now stable and generally available. Plus, get a preview of Gemini 2.5 Flash-Lite, our fastest + most cost-efficient 2.5 model yet. 🔦 Exciting steps as we expand our 2.5 series of hybrid reasoning models that deliver amazing performance at the

Gemini 2.5 Pro + 2.5 Flash are now stable and generally available. Plus, get a preview of Gemini 2.5 Flash-Lite, our fastest + most cost-efficient 2.5 model yet. 🔦

Exciting steps as we expand our 2.5 series of hybrid reasoning models that deliver amazing performance at the
Oriol Vinyals (@oriolvinyalsml) 's Twitter Profile Photo

Hello Gemini 2.5 Flash-Lite! So fast, it codes *each screen* on the fly (Neural OS concept 👇). The frontier isn't always about large models and beating benchmarks. In this case, a super fast & good model can unlock drastic use cases. Read more: blog.google/products/gemin…

jack morris (@jxmnop) 's Twitter Profile Photo

NEW RESEARCH: Approximating Language Model Training Data from Weights ever wonder how much information is available in an open-weights model? DeepSeek R1 weights are 1.2 TB... what can we learn from all those bits? our method reverses LLM finetuning to recover data: 🧵

NEW RESEARCH:   Approximating Language Model Training Data from Weights

ever wonder how much information is available in an open-weights model?  

DeepSeek R1 weights are 1.2 TB... 

what can we learn from all those bits?

our method reverses LLM finetuning to recover data: 🧵
Kimi.ai (@kimi_moonshot) 's Twitter Profile Photo

Meet Kimi-Researcher - an autonomous agent that excels at multi-turn search and reasoning. Powered by k 1.5 and trained with end-to-end agentic RL. Achieved 26.9% pass@1 on Humanity's Last Exam, 69% pass@1 on xbench. 🔗 Tech blog:moonshotai.github.io/Kimi-Researche…

Meet Kimi-Researcher - an autonomous agent that excels at multi-turn search and reasoning. Powered by k 1.5 and trained with end-to-end agentic RL. 

Achieved 26.9% pass@1 on Humanity's Last Exam, 69% pass@1 on xbench.

🔗 Tech blog:moonshotai.github.io/Kimi-Researche…
Chris Donahue (@chrisdonahuey) 's Twitter Profile Photo

Excited to announce 🎵Magenta RealTime, the first open weights music generation model capable of real-time audio generation with real-time control. 👋 **Try Magenta RT on Colab TPUs**: colab.research.google.com/github/magenta… 👀 Blog post: g.co/magenta/rt 🧵 below

Richard Socher (@richardsocher) 's Twitter Profile Photo

If you studied algorithms, I'm sure you've heard of Dijkstra’s algorithm to find the shortest paths between nodes in a weighted graph. Super useful in scenarios such as road networks, where it can determine the shortest route from a starting point to various destinations. It's

If you studied algorithms, I'm sure you've heard of Dijkstra’s algorithm to find the shortest paths between nodes in a weighted graph. Super useful in scenarios such as road networks, where it can determine the shortest route from a starting point to various destinations. It's
Tu Vu (@tuvllms) 's Twitter Profile Photo

One appealing property of Seal-0 is that you need to search to achieve better performance, but the more you search, the more conflicting evidence you encounter. This poses a major challenge for deep research agents / frontier LLMs. Congrats Kimi.ai on the strong results!

Google DeepMind (@googledeepmind) 's Twitter Profile Photo

We’re bringing powerful AI directly onto robots with Gemini Robotics On-Device. 🤖 It’s our first vision-language-action model to help make robots faster, highly efficient, and adaptable to new tasks and environments - without needing a constant internet connection. 🧵

Andrej Karpathy (@karpathy) 's Twitter Profile Photo

+1 for "context engineering" over "prompt engineering". People associate prompts with short task descriptions you'd give an LLM in your day-to-day use. When in every industrial-strength LLM app, context engineering is the delicate art and science of filling the context window

Demis Hassabis (@demishassabis) 's Twitter Profile Photo

Thrilled to introduce AlphaGenome, our new DNA sequence model now available via our AlphaGenome API. Really excited to see how the scientific community uses AlphaGenome’s predictions to understand genome function, drive biological discoveries, develop new treatments, and more...

Tu Vu (@tuvllms) 's Twitter Profile Photo

Excited to share that our paper on model merging at scale has been accepted to Transactions on Machine Learning Research (TMLR). Huge congrats to my intern Prateek Yadav and our awesome co-authors Jonathan Lai, Alexandra Chronopoulou, Manaal Faruqui, Mohit Bansal, and Tsendsuren 🎉!!

Excited to share that our paper on model merging at scale has been accepted to Transactions on Machine Learning Research (TMLR). Huge congrats to my intern <a href="/prateeky2806/">Prateek Yadav</a> and our awesome co-authors <a href="/_JLai/">Jonathan Lai</a>, <a href="/alexandraxron/">Alexandra Chronopoulou</a>, <a href="/manaalfar/">Manaal Faruqui</a>, <a href="/mohitban47/">Mohit Bansal</a>, and <a href="/TsendeeMTS/">Tsendsuren</a> 🎉!!