Jiayuan Mao (@maojiayuan) 's Twitter Profile
Jiayuan Mao

@maojiayuan

PhD Student at @MIT_LISLab/@MITCoCoSci

ID: 1054159620

linkhttp://jiayuanm.com calendar_today02-01-2013 04:18:41

21 Tweet

1,1K Followers

163 Following

Zhutian Yang (@zhutianyang_) 's Twitter Profile Photo

Our #CoRL2023 paper shows that by composing the energies of diffusion models trained to sample for individual constraint types such as collision-free, spatial relations, and physical stability, it can solve novel combinations of known constraints. diffusion-ccsp.github.io (🧵 1/N)

Yilun Du (@du_yilun) 's Twitter Profile Photo

Introducing a way to convert synthesized robot videos to robot execution without using any action labels! flow-diffusion.github.io We also release a codebase (with pretrained models) for text-to-video generation. Train your own models for robot control in only 1 day with 4 GPUs!

Joy Hsu (@joycjhsu) 's Twitter Profile Photo

What’s left w/ foundation models? We found that they still can't ground modular concepts across domains. We present Logic-Enhanced FMs:🤝FMs & neuro-symbolic concept learners. We learn abstractions of concepts like “left” across domains & do domain-independent reasoning w/ LLMs.

Nishanth Kumar (@nishanthkumar23) 's Twitter Profile Photo

Ever heard about "Bilevel Planning" or "Task and Motion Planning", but been unsure what those words mean? Ever wanted a gentle intro to these methods so you can just understand what's going on? Our new blog post might help! lis.csail.mit.edu/bilevel-planni…

Jiayuan Mao (@maojiayuan) 's Twitter Profile Photo

Definitely one of my top 3 favourite papers :) It marries deep learning with a minimal set of universal grammar rules for grounded language learning. It draws inspiration from lexicalist linguistics and cognitive science (bootstrapping from core knowledge).

Jiayuan Mao (@maojiayuan) 's Twitter Profile Photo

Check out our new framework that automatically generates planning domain knowledge with LLMs, learns to ground it, and verifies it through interaction. We believe that learning such verifiable and compositional planning representations from language is important for embodied AI!

Fangchen Liu (@fangchenliu_) 's Twitter Profile Photo

Can we leverage VLMs for robot manipulation in the open world? Checkout our new work MOKA, a simple and effective visual prompting method!

Fangchen Liu (@fangchenliu_) 's Twitter Profile Photo

The key idea is to query GPT-4V to perform multiple choice from a set of keypoints and waypoints. Here is an example: suppose the current task is to sweep the trash bag off the table. We mark the involved objects with a set of points, and overlay the image with grids.

The key idea is to query GPT-4V to perform multiple choice from a set of keypoints and waypoints. Here is an example: suppose the current task is to sweep the trash bag off the table. We mark the involved objects with a set of points, and overlay the image with grids.
Chen Wang (@chenwang_j) 's Twitter Profile Photo

Can we use wearable devices to collect robot data without actual robots? Yes! With a pair of gloves🧤! Introducing DexCap, a portable hand motion capture system that collects 3D data (point cloud + finger motion) for training robots with dexterous hands Everything open-sourced

Yilun Du (@du_yilun) 's Twitter Profile Photo

Introducing our @icml_conf paper: Learning Iterative Reasoning through Energy Diffusion! We formulate reasoning as optimizing a sequence of energy landscapes. This enables us to solve harder problems at test time with more complex optimization. Website: energy-based-model.github.io/ired/

Joy Hsu (@joycjhsu) 's Twitter Profile Photo

What makes a maze look like a maze? Humans can reason about infinitely many instantiations of mazes—made of candy canes, sticks, icing, yarn, etc. But VLMs often struggle to make sense of such visual abstractions. We improve VLMs' ability to interpret these abstract concepts.

Jiayuan Mao (@maojiayuan) 's Twitter Profile Photo

Excited to announce our upcoming workshop: “Planning in the Era of LLMs”! 🌐🤖 Join us as we explore the intersection of planning representations, algorithms, and the power of LLMs. Let’s discuss how combining these insights and solutions can unlock new capabilities in AI.

Joy Hsu (@joycjhsu) 's Twitter Profile Photo

Excited to bring back the 2nd Workshop on Visual Concepts at #CVPR2025 2025, this time with a call for papers! We welcome submissions on the following topics. See our website for more info: sites.google.com/stanford.edu/w… Join us & a fantastic lineup of speakers in Tennessee!

Excited to bring back the 2nd Workshop on Visual Concepts at <a href="/CVPR/">#CVPR2025</a> 2025, this time with a call for papers!

We welcome submissions on the following topics. See our website for more info:
sites.google.com/stanford.edu/w…

Join us &amp; a fantastic lineup of speakers in Tennessee!
Haonan Chen (@haonanchen_) 's Twitter Profile Photo

Excited to organize Workshop on Learning Meets Model-Based Methods for Contact-Rich Manipulation @ ICRA 2025! We welcome submissions on a range of topics—check out our website for details: contact-rich.github.io Join us for an incredible lineup of speakers! #ICRA2025

Excited to organize Workshop on Learning Meets Model-Based Methods for Contact-Rich Manipulation  @ ICRA 2025!

We welcome submissions on a range of topics—check out our website for details:
 contact-rich.github.io

Join us for an incredible lineup of speakers! #ICRA2025
Manling Li (@manlingli_) 's Twitter Profile Photo

Tutorial on "Foundation Models Meet Embodied Agents", with Yunzhu Li Jiayuan Mao Wenlong Huang . We categorize LLMs, VLMs, VLAs and their usages under a MDP formulation. Come and talk with us! Also on Zoom! …models-meet-embodied-agents.github.io Feb 25 8:30-12:30 Location: 118A room,

Tutorial on "Foundation Models Meet Embodied Agents", with <a href="/YunzhuLiYZ/">Yunzhu Li</a> <a href="/maojiayuan/">Jiayuan Mao</a> <a href="/wenlong_huang/">Wenlong Huang</a> . 

We categorize LLMs, VLMs, VLAs and their usages under a MDP formulation.

Come and talk with us! Also on Zoom!

…models-meet-embodied-agents.github.io
Feb 25 8:30-12:30
Location: 118A room,
Shao-Hua Sun (@shaohua0116) 's Twitter Profile Photo

We will organize an #ICML2025 Workshop on Programmatic Representations for Agent Learning, bringing together experts in decision-making and code generation to explore how structured representations can make agent learning more interpretable, generalizable, efficient, and safe!

We will organize an #ICML2025 Workshop on Programmatic Representations for Agent Learning, bringing together experts in decision-making and code generation to explore how structured representations can make agent learning more interpretable, generalizable, efficient, and safe!
Jiayuan Mao (@maojiayuan) 's Twitter Profile Photo

Excited to see more work at the intersection of programmatic representations, program synthesis, and learning! If you’re exploring programmatic representations for agents, consider submitting to our ICML & RLC workshops!