Huaxiu Yao✈️ICLR 2025🇸🇬 (@huaxiuyaoml) 's Twitter Profile
Huaxiu Yao✈️ICLR 2025🇸🇬

@huaxiuyaoml

Assistant Professor of Computer Science @UNC @unccs @uncsdss | Postdoc @StanfordAILab | Ph.D. @PennState | #foundationmodels, #EmbodiedAI, #Alignment | he/him

ID: 1411871581

linkhttp://www.huaxiuyao.io/ calendar_today08-05-2013 03:36:31

418 Tweet

3,3K Followers

566 Following

Chelsea Finn (@chelseabfinn) 's Twitter Profile Photo

Can we prompt robots, just like we prompt language models? With hierarchy of VLA models + LLM-generated data, robots can: - reason through long-horizon tasks - respond to variety of prompts - handle situated corrections Blog post & paper: pi.website/research/hirob…

Mohit Bansal (@mohitban47) 's Twitter Profile Photo

Thank you AAAI for the honor & the fun ceremonies -- humbled to be inducted as a AAAI Fellow in esteemed company 🙏 PS. I am still around today in Philadelphia if anyone wants to meet up at #AAAI2025 :-) Thanks once again to everyone (students+postdocs+collaborators,

Thank you <a href="/RealAAAI/">AAAI</a> for the honor &amp; the fun ceremonies -- humbled to be inducted as a AAAI Fellow in esteemed company 🙏 
PS. I am still around today in Philadelphia if anyone wants to meet up at #AAAI2025 :-)

Thanks once again to everyone (students+postdocs+collaborators,
Prithvijit (@prithvijitch) 's Twitter Profile Photo

Join us at the WorldModelBench workshop at #CVPR2025 where we'll tackle systematic evaluation of World Models! Focus: benchmarks, metrics, downstream tasks, and safety. Submit papers now: worldmodelbench.github.io

No Priors (@nopriorspod) 's Twitter Profile Photo

Are we at a turning point in robotics? New interview with Chelsea Finn founder of Physical Intelligence: we talk about pi’s approach to robotics foundation models, generalization, data generation, humanoids, and comparisons to self-driving. links 👇

Are we at a turning point in robotics?

New interview with <a href="/chelseabfinn/">Chelsea Finn</a> founder of <a href="/physical_int/">Physical Intelligence</a>: we talk about pi’s approach to robotics foundation models, generalization, data generation, humanoids, and comparisons to self-driving. links 👇
Peng (Richard) Xia ✈️ ICLR 2025 (@richardxp888) 's Twitter Profile Photo

🚀 Introducing MDocAgent! 🧐📄 📚 Ever struggled with AI that can’t handle complex documents filled with text, images, tables, and figures? 💡 Enter MDocAgent 🧠🤖—a next-gen multi-modal multi-agent framework that revolutionizes document understanding! #AI #DocQA #LLM #Agent

🚀 Introducing MDocAgent! 🧐📄

📚 Ever struggled with AI that can’t handle complex documents filled with text, images, tables, and figures?

💡 Enter MDocAgent 🧠🤖—a next-gen multi-modal multi-agent framework that revolutionizes document understanding!

#AI #DocQA #LLM #Agent
Huaxiu Yao✈️ICLR 2025🇸🇬 (@huaxiuyaoml) 's Twitter Profile Photo

🚀 Launched MDocAgent—a multi-agent framework that crushes complex documents (text, images, tables!) with specialized AI agents. Outperforms SOTA by 12.1%! See details in Peng (Richard) Xia's thread.

Huaxiu Yao✈️ICLR 2025🇸🇬 (@huaxiuyaoml) 's Twitter Profile Photo

I’ll be at RIKEN Center for Advanced Intelligence Project 🇯🇵 (Apr 19–21) and #ICLR2025 🇸🇬 (Apr 22–28), presenting our latest ICLR work and organizing the FM in the wild workshop. If you’d like to chat about (1) VLM / VLA / AI agents or (2) PhD openings in my lab, DM or email me—let’s connect! My students

I’ll be at <a href="/RIKEN_AIP_EN/">RIKEN Center for Advanced Intelligence Project</a> 🇯🇵 (Apr 19–21) and #ICLR2025 🇸🇬 (Apr 22–28), presenting our latest ICLR work and organizing the FM in the wild workshop. 

If you’d like to chat about (1) VLM / VLA / AI agents or (2) PhD openings in my lab, DM or email me—let’s connect!

My students
Sergey Levine (@svlevine) 's Twitter Profile Photo

π-0.5 is here, and it can generalize to new homes! Some fun experiments with my colleagues at Physical Intelligence, introducing π-0.5 (“pi oh five”). Our new VLA can put dishes in the sink, clean up spills and do all this in homes that it was not trained in🧵👇

LLM360 (@llm360) 's Twitter Profile Photo

The MBZUAI IFM and the LLM360 team's first day at ICLR 2026, come to visit our new Institute of Foundation Models! Booth D04 in Hall 2! We’re looking forward to meeting researchers and engineers to introduce them to MBZUAI .

The MBZUAI IFM and the LLM360 team's first day at <a href="/iclr_conf/">ICLR 2026</a>, come to visit our new Institute of Foundation Models!   Booth D04 in Hall 2!

We’re looking forward to meeting researchers and engineers to introduce them to <a href="/mbzuai/">MBZUAI</a> .
Chelsea Finn (@chelseabfinn) 's Twitter Profile Photo

I'm giving two talks today/Sunday at #ICLR2025 ! - Post-Training Robot Foundation Models (Robot Learning Workshop @ 12:50 pm) - Robot Foundation Models with Open-Ended Generalization (Foundation Models in the Wild @ 2:30 pm) Will cover π-0, Demo-SCORE, Hi Robot, & π-0.5.

Qwen (@alibaba_qwen) 's Twitter Profile Photo

Introducing Qwen3! We release and open-weight Qwen3, our latest large language models, including 2 MoE models and 6 dense models, ranging from 0.6B to 235B. Our flagship model, Qwen3-235B-A22B, achieves competitive results in benchmark evaluations of coding, math, general

Introducing Qwen3! 

We release and open-weight Qwen3, our latest large language models, including 2 MoE models and 6 dense models, ranging from 0.6B to 235B. Our flagship model, Qwen3-235B-A22B, achieves competitive results in benchmark evaluations of coding, math, general
Fan Nie (@fannie1208) 's Twitter Profile Photo

🚀 Excited to share that “#FactTest: Factuality Testing in Large Language Models with Finite-Sample and Distribution-Free Guarantees” has been accepted to #ICML25 ! 🎉 📄 Paper: arxiv.org/abs/2411.02603 💻 Code: github.com/fannie1208/Fac… 1/N

🚀 Excited to share that “#FactTest: Factuality Testing in Large Language Models with Finite-Sample and Distribution-Free Guarantees” has been accepted to #ICML25 ! 🎉

📄 Paper: arxiv.org/abs/2411.02603
💻 Code: github.com/fannie1208/Fac…
1/N
Joel Jang (@jang_yoel) 's Twitter Profile Photo

Introducing 𝐃𝐫𝐞𝐚𝐦𝐆𝐞𝐧! We got humanoid robots to perform totally new 𝑣𝑒𝑟𝑏𝑠 in new environments through video world models. We believe video world models will solve the data problem in robotics. Bringing the paradigm of scaling human hours to GPU hours. Quick 🧵

Yiyang Zhou (@aiyiyangz) 's Twitter Profile Photo

🔥 ReAgent-V Released! 🔥 A unified video framework with reflection and reward-driven optimization. ✨ Real-time self-correction. ✨ Triple-view reflection. ✨ Auto-selects high-reward samples for training.

🔥 ReAgent-V Released! 🔥

A unified video framework with reflection and reward-driven optimization.

✨ Real-time self-correction.
✨ Triple-view reflection.
✨ Auto-selects high-reward samples for training.
Peng (Richard) Xia ✈️ ICLR 2025 (@richardxp888) 's Twitter Profile Photo

🚑 Introducing MedAgentRL – a dynamic, RL-powered multi-agent framework for medical multimodal reasoning! 🤝🧠 Tired of AI models that fumble across specialties? MedAgentRL boosts performance by 20% over SFT, delivering smarter, collaborative diagnosis. 🏥 #MedAI #LLM #Agent #RL

🚑 Introducing MedAgentRL – a dynamic, RL-powered multi-agent framework for medical multimodal reasoning! 🤝🧠 Tired of AI models that fumble across specialties? MedAgentRL boosts performance by 20% over SFT, delivering smarter, collaborative diagnosis. 🏥

#MedAI #LLM #Agent #RL
Huaxiu Yao✈️ICLR 2025🇸🇬 (@huaxiuyaoml) 's Twitter Profile Photo

🎓 Meet EduVisAgent: the next-gen AI teaching assistant for STEM education. We're shifting from static answers to dynamic, visual-guided reasoning. 📊 We introduce EduVisBench, the first benchmark for visual teaching quality with 1,154 STEM questions, and EduVisAgent, a