Katerina Fragkiadaki (@katerinafragiad) 's Twitter Profile
Katerina Fragkiadaki

@katerinafragiad

Associate Professor @CMU working on #AI #ComputerVision #Robotics #LanguageGrounding

ID: 1681924306299650049

linkhttps://www.cs.cmu.edu/~katef/ calendar_today20-07-2023 07:10:19

15 Tweet

667 Followers

59 Following

Katerina Fragkiadaki (@katerinafragiad) 's Twitter Profile Photo

Come join us in our CoRL workshop on generalists agents! How can we build robot generalists using learning from video, scaling up simulators, real robot exploration, generative AI and more? #FirstTweet

Katerina Fragkiadaki (@katerinafragiad) 's Twitter Profile Photo

Gabe's work on memory-augmented prompting of LLMs supports open-ended semantic parsing and task planning without forgetting. The model can store, retrieve and adapt instructed plans on-the-fly at deployment time!

Pushkal Katara (@pushkalkatara) 's Twitter Profile Photo

🤖🌐How to scale up data across diverse tasks and environments for robot skill learning? Harness the power of language, vision generative models in simulation! Excited to share Gen2Sim, a step towards autonomous robotic skill acquisition in simulation. gen2sim.github.io

Katerina Fragkiadaki (@katerinafragiad) 's Twitter Profile Photo

Check out our latest work on 3D feature field transformers for learning manipulation policies from demonstrations, with SOTA performance in RLBench.

Katerina Fragkiadaki (@katerinafragiad) 's Twitter Profile Photo

With ODIN, 3D perception most benefits from 2D feature pre-training, and is used in the real world, outside dataset-given 3D meshes. Congrats Ayush and ODIN team!

Katerina Fragkiadaki (@katerinafragiad) 's Twitter Profile Photo

Very excited to speak tomorrow on unifying 2D/3D models of images, language and actions at multimodalitiesfor3dscenes.github.io. On Tuesday, I will talk about generative video perception at generative-vision.github.io/workshop-CVPR-… and memory-prompted 3D parsing at 3dcompat-dataset.org/workshop/C3DV2…. See you #CVPR2025 !

Katerina Fragkiadaki (@katerinafragiad) 's Twitter Profile Photo

Come chat with us today about Diffusion-ES, that combines evolutionary search with diffusion models for efficient planning! Code is also now publicly available.

Katerina Fragkiadaki (@katerinafragiad) 's Twitter Profile Photo

We show an LLM/VLM agent can transform plans into plan abstractions by adding language comments on preconditions, state changes and subgoals, using VLM's knowledge and human feedback. In-context planning with retrieved abstractions sets a new SOTA across domains.

Katerina Fragkiadaki (@katerinafragiad) 's Twitter Profile Photo

The Genesis physics engine is just released, incredible effort from Xian and his team. It is a most general, fast and easy to use physics engine, which we hope it will accelerate robotics research and beyond.

Katerina Fragkiadaki (@katerinafragiad) 's Twitter Profile Photo

A thorough apples-to-apples speed comparison between IsaacGym, MujocoMJX and GENESIS in static and dynamic scenes, with and without collisions. We hope to have a paper ready within the next few months.

Katerina Fragkiadaki (@katerinafragiad) 's Twitter Profile Photo

We train UniDisc - a diffusion model that can generate image and text jointly! Using diffusion instead of AR permits high controllability including infilling in multimodal space, not possible with any other multimodal generative model currently. Congrats to Alex and Mihir!

Katerina Fragkiadaki (@katerinafragiad) 's Twitter Profile Photo

When data is limited and training spans multiple epochs, discrete diffusion beats autoregressive models for text generation. Check out Mihir and Mengning’s thorough analysis: