Yiming Dou (@_yimingdou) 's Twitter Profile
Yiming Dou

@_yimingdou

Ph.D. student at UMich | B.Eng. from SJTU | Previous intern at Stanford | Computer Vision, Multimodal, Robotics

ID: 1507890461398253575

linkhttps://dou-yiming.github.io/ calendar_today27-03-2022 01:21:26

64 Tweet

695 Followers

867 Following

Zichen Wang (@zichen2501) 's Twitter Profile Photo

Differentiable rendering made SIMPLE❗️ Differentiating physically based renderers is hard: Dirac-delta discontinuities arise at object silhouette. Our #SIGGRAPHAsia2024 work shows how a simple relaxation can rescue the day, enabling easy 3D reconstruction and relighting! (1/N)

Ayush Shrivastava (@ayshrv) 's Twitter Profile Photo

We present Global Matching Random Walks, a simple self-supervised approach to the Tracking Any Point (TAP) problem, accepted to #ECCV2024. We train a global matching transformer to find cycle consistent tracks through video via contrastive random walks (CRW).

Junyi Zhang (@junyi42) 's Twitter Profile Photo

Excited to share MonST3R! -- a simple way to estimate geometry from unposed video of dynamic scene We achieve competitive results on several downstreams (video depth, camera pose) and believe this is a promising step toward feed-forward 4D reconstruction monst3r-project.github.io

Daniel Geng (@dangengdg) 's Twitter Profile Photo

What happens when you train a video generation model to be conditioned on motion? Turns out you can perform "motion prompting," just like you might prompt an LLM! Doing so enables many different capabilities. Here’s a few examples – check out this thread 🧵 for more results!

Yuanchen_Ju (@ju_yuanchen) 's Twitter Profile Photo

🍌We present DenseMatcher! 🤖️DenseMatcher enables robots to acquire generalizable skills across diverse object categories by only seeing one demo, by finding correspondences between 3D objects even with different types, shapes, and appearances.

Sarah Jabbour (@sarahjabbour_) 's Twitter Profile Photo

I’m on the PhD internship market for Spr/Summer 2025! I have experience in multimodal AI (EHR, X-ray, text), explainability for image models w/ genAI, clinician-AI interaction (surveyed 700+ doctors), and tabular foundation models. Please reach out if you think there’s a fit!

Yuanchen_Ju (@ju_yuanchen) 's Twitter Profile Photo

🧩#CVPR2025🌷Introducing Two By Two✌️: The First Large-Scale Daily Pairwise Assembly Dataset with SE(3)-Equivariant Pose Estimation. 🤖2BY2 helps robots master daily 3D assembly tasks—like plugging sockets or arranging flowers—across diverse objects! 🐨Co-lead by Yu Qi

Chris Rockwell (@_crockwell) 's Twitter Profile Photo

Ever wish YouTube had 3D labels? 🚀Introducing🎥DynPose-100K🎥, an Internet-scale collection of diverse videos annotated with camera pose! Applications include camera-controlled video generation🤩and learned dynamic pose estimation😯 Download: huggingface.co/datasets/nvidi…

Daniel Geng (@dangengdg) 's Twitter Profile Photo

Hello! If you like pretty images and videos and want a rec for CVPR oral session, you should def go to Image/Video Gen, Friday at 9am: I'll be presenting "Motion Prompting" Ryan Burgert will be presenting "Go with the Flow" and Pascal CHANG will be presenting "LookingGlass"

Jeongsoo Park (@jespark0) 's Twitter Profile Photo

Can AI image detectors keep up with new fakes? Mostly, no. Existing detectors are trained using a handful of models. But there are thousands in the wild! Our work, Community Forensics, uses 4800+ generators to train detectors that generalize to new fakes. #CVPR2025 🧵 (1/5)

Ayush Shrivastava (@ayshrv) 's Twitter Profile Photo

Excited to share our CVPR 2025 paper on cross-modal space-time correspondence! We present a method to match pixels across different modalities (RGB-Depth, RGB-Thermal, Photo-Sketch, and cross-style images) — trained entirely using unpaired data and self-supervision. Our

Excited to share our CVPR 2025 paper on cross-modal space-time correspondence!

We present a method to match pixels across different modalities (RGB-Depth, RGB-Thermal, Photo-Sketch, and cross-style images) — trained entirely using unpaired data and self-supervision.

Our
Linyi Jin (@jin_linyi) 's Twitter Profile Photo

Hello! If you are interested in dynamic 3D or 4D, don't miss the oral session 3A at 9 am on Saturday: Zhengqi Li will be presenting "MegaSaM" I'll be presenting "Stereo4D" and Qianqian Wang will be presenting "CUT3R"

Paul Liang (@pliang279) 's Twitter Profile Photo

Despite much progress in AI, the ability for AI to 'smell' like humans remains elusive. Smell AIs 🤖👃can be used for allergen sensing (e.g., peanuts or gluten in food), hormone detection for health, safety & environmental monitoring, quality control in manufacturing, and more.