Xavier Puig @ ICLR (@xavierpuigf) 's Twitter Profile
Xavier Puig @ ICLR

@xavierpuigf

Research Scientist at FAIR @AIatMeta working on EmbodiedAI | PhD @MIT_CSAIL

ID: 383903725

linkhttp://xavierpuigf.com calendar_today02-10-2011 18:32:10

280 Tweet

1,1K Followers

883 Following

AI at Meta (@aiatmeta) 's Twitter Profile Photo

Additionally, looking towards the future, weโ€™re releasing PARTNR: a benchmark for Planning And Reasoning Tasks in humaN-Robot collaboration. Built on Habitat 3.0, itโ€™s the largest benchmark of its kind to study and evaluate human-robot collaboration in household activities By

Manling Li (@manlingli_) 's Twitter Profile Photo

[NeurIPS D&B Oral] Embodied Agent Interface: Benchmarking LLMs for Embodied Agents A single line of code to evaluate your model! ๐ŸŒŸStandardize Goal Specifications: LTL ๐ŸŒŸStandardize Modules and Interfaces: 4 modules, 438 tasks, 1475 goals ๐ŸŒŸStandardize Fine-grained Metrics: 18

Jiaman Li (@jiaman01) 's Twitter Profile Photo

๐Ÿค– Introducing Human-Object Interaction from Human-Level Instructions! First complete system that generates physically plausible, long-horizon human-object interactions with finger motions in contextual environments, driven by human-level instructions. ๐Ÿ” Our approach: - LLMs

Tianyu Li EasyPaperSniper (@sniperpaper) 's Twitter Profile Photo

The trained policy can be integrated with a high-level planner for real-world applications. By combining our object manipulation policy with user commands, we demonstrate its effectiveness in real-world scenariosโ€”such as moving large trash carts. (6/8)

Xavier Puig @ ICLR (@xavierpuigf) 's Twitter Profile Photo

๐Ÿช‘How do you train robots to move furniture? This requires robots to synchronize whole-body movements, making teleoperation or RL approaches challenging. Check out this amazing work by Tianyu Li EasyPaperSniper, using human demonstrations to train robots to move furniture in the real world!

AI at Meta (@aiatmeta) 's Twitter Profile Photo

Meta PARTNR is a benchmark for planning and reasoning in embodied multi-agent tasks. This large-scale human and robot collaboration benchmark was core to our recent demos and also informs our work as scientists and engineers pushing this field of study forward.

Chuanyang Jin (@chuanyang_jin) 's Twitter Profile Photo

How to achieve human-level open-ended machine Theory of Mind? Introducing #AutoToM: a fully automated and open-ended ToM reasoning method combining the flexibility of LLMs with the robustness of Bayesian inverse planning, achieving SOTA results across five benchmarks. ๐Ÿงต[1/n]

How to achieve human-level open-ended machine Theory of Mind?

Introducing #AutoToM: a fully automated and open-ended ToM reasoning method combining the flexibility of LLMs with the robustness of Bayesian inverse planning, achieving SOTA results across five benchmarks. ๐Ÿงต[1/n]
Ram Ramrakhya (@ramramrakhya) 's Twitter Profile Photo

๐ŸšจNew Preprint ๐Ÿšจ Embodied agents ๐Ÿค– operating in indoor environments must interpret ambiguous and under-specified human instructions. A capable household robot ๐Ÿค– should recognize ambiguity and ask relevant clarification questions to infer the user๐Ÿง‘โ€๐Ÿš’ intent accurately, leading

๐ŸšจNew Preprint ๐Ÿšจ

Embodied agents ๐Ÿค– operating in indoor environments must interpret ambiguous and under-specified human instructions. A capable household robot ๐Ÿค– should recognize ambiguity and ask relevant clarification questions to infer the user๐Ÿง‘โ€๐Ÿš’ intent accurately, leading
Xavier Puig @ ICLR (@xavierpuigf) 's Twitter Profile Photo

How do we enable agents to perform tasks even when these are underspecified? In this work, led by Ram Ramrakhya, we train VLA agents via RL to decide when to act in the environment or ask clarifying questions, enabling them to handle ambiguous instructions ram81.github.io/projects/ask-tโ€ฆ

Xavier Puig @ ICLR (@xavierpuigf) 's Twitter Profile Photo

I will be at ICLR to present PARTNR. Reach out if you want to talk about our work at FAIR or interesting problems in Robotics!

Mandi Zhao (@zhaomandi) 's Twitter Profile Photo

DexMachina lets us perform a functional comparison between different dexterous hands: we evaluate 6 hands on 4 challenging long-horizon tasks, and found that larger, fully actuated hands learn better and faster, and high DoF is more important than having human-like hand sizes โ€“

DexMachina lets us perform a functional comparison between different dexterous hands: we evaluate 6 hands on 4 challenging long-horizon tasks, and found that larger, fully actuated hands learn better and faster, and high DoF is more important than having human-like hand sizes โ€“
Roozbeh Mottaghi (@roozbehmottaghi) 's Twitter Profile Photo

I'll be giving two talks at the #CVPR2025 workshops: 3D LLM/VLA 3d-llm-vla.github.io and POETS poets2024.github.io/poets2025/. ๐Ÿงต

Xavier Puig @ ICLR (@xavierpuigf) 's Twitter Profile Photo

I will be talking at the #CVPR2025 workshop on Humanoid Agents, tomorrow June 11th at 9:30 am. I will discuss how humanoid agents can help us improve human-robot collaboration. See you there! humanoid-agents.github.io

I will be talking at the #CVPR2025 workshop on Humanoid Agents, tomorrow June 11th at 9:30 am. I will discuss how humanoid agents can help us improve human-robot collaboration. See you there!
humanoid-agents.github.io
Tianmin Shu (@tianminshu) 's Twitter Profile Photo

๐Ÿš€ Excited to introduce SimWorld: an embodied simulator for infinite photorealistic world generation ๐Ÿ™๏ธ populated with diverse agents ๐Ÿค– If you are at #CVPR2025, come check out the live demo ๐Ÿ‘‡ Jun 14, 12:00-1:00 pm at JHU booth, ExHall B Jun 15, 10:30 am-12:30 pm, #7, ExHall B

Yixuan Wang (@yxwangbot) 's Twitter Profile Photo

๐Ÿค– Does VLA models really listen to language instructions? Maybe not ๐Ÿ‘€ ๐Ÿš€ Introducing our RSS paper: CodeDiffuser -- using VLM-generated code to bridge the gap between **high-level language** and **low-level visuomotor policy** ๐ŸŽฎ Try the live demo: robopil.github.io/code-diffuser/ (1/9)

Xavier Puig @ ICLR (@xavierpuigf) 's Twitter Profile Photo

Check out our workshop on Continual Robot Learning from Humans, at #RSS2025, with amazing speakers covering topics including learning from human visual demonstrations, generative models for continual robot learning or the role of LLMs in embodied contexts โ€ฆ-robot-learning-from-humans.github.io