[1/10] 🚨 We present our recent Snap Inc. project: Stable Flow --- A training-free method that performs various types of image editing operations (e.g., non-rigid editing, object addition and replacement) using flow models.
Project page: omriavrahami.com/stable-flow
🚀 Excited to share ObjectMate, our latest Google AI project! A new approach to zero-shot subject-driven generation and object insertion.
🔗 Explore our project page: object-mate.com
📄 Arxiv: arxiv.org/abs/2412.08645
I'm excited about the recent feature we've been working on at Pika - action editing! 🏃 📹
Video editing methods often focus on spatial edits. Our method can directly edit the *action* of objects, even when the video already contains a distinctive motion (e.g, 👋).
Many modern SpeechLMs are trained with Speech-Text interleaving. How does this impact scaling trends?
In our new paper, we train several dozen SLMs, and show - quite a lot! So there is room for optimism 😊
Key insights, code, models, full paper 👇🏻
Our work maps hereditary relationships between models. We find that weights🏋️ are sufficient for decoding the Origin of Models🌳
Presenting today at ICLR 2026, 15:00–17:30, Hall 3, poster #360.
Come by to see our method, visualizations, and interactive demo atlas👀🚀 #ICLR2025
🎵💬 If you are interested in Audio Tokenisers, you should check out our new work!
We empirically analysed existing tokenisers from every way - reconstruction, downstream, LMs and more.
Grab yourself a ☕/🍺 and sit down for a read!