
Ying Fan
@yingfan_bot
CS PhD candidate @UWMadison | Ex-intern @GoogleAI @MSFTResearch
ID: 850996562373357568
http://yingfan-bot.github.io 09-04-2017 08:59:26
42 Tweet
284 Followers
127 Following




Why are diffusion models so good? Our NeurIPS work by Dohyun Kwon , Ying Fan and Kangwook Lee presents a plausible explanation for it. 🧵

How to build *TruthGPT*? I listened to a talk by the legendary John Schulman. It's densely packed with lots of deep insight. Key takeaways: - Supervised finetuning (or behavior cloning) makes the model prone to hallucination, while RL mitigates it. - NLP is far from done! 1/🧵



🔥Check out our ICML Conference ICML23' work on training diffusion models with policy gradient for shortcuts, which is the first work to use RL for training diffusion models to our knowledge. Check out our Arxiv paper arxiv.org/abs/2301.13362 & an exciting follow-up work coming soon!










Huge thanks to UMD Center for Machine Learning for the recognition, and grateful to Kangwook Lee for being an amazing advisor!

While I couldn't make it to #NeurIPS2024 this time, Ching-An Cheng and Aditya Modi will present our work on offline contextual goal-oriented RL @ West Ballroom A-D #6206 on Thu (Poster Session 3 West)! Also check our paper here: arxiv.org/abs/2408.07753.


