Rujun Han (@hanrujun) 's Twitter Profile
Rujun Han

@hanrujun

Research Scientist @Google working on RAG, LLM evaluation, post-training, and alignment. NLP+ML Ph.D. from @USC_ISI. Ex-@AWS AI, ex-@federalreserve.

ID: 1131657545872027649

linkhttps://rujunhan.github.io calendar_today23-05-2019 20:25:58

39 Tweet

299 Followers

204 Following

Rujun Han (@hanrujun) 's Twitter Profile Photo

Ready to take knowledge distillation for LLMs to the next level? Check out our Speculative Knowledge Distillation paper which leverages samples composed of the best student and teacher tokens to achieve SOTA results. Collaboration with our wonderful intern Wenda Xu!

Rujun Han (@hanrujun) 's Twitter Profile Photo

Another #EMNLP2024 paper from the collaboration with my previous colleagues at AWS AI. Please join our oral presentation to learn more about how to mitigate the trade-off between LLM instruction following and grounding.

Violet Peng (@violetnpeng) 's Twitter Profile Photo

This Thanksgiving, Iโ€™m deeply grateful for opportunity to run for NAACL Board Member alongside so many incredible candidates who share a passion for making a difference! I hope to earn your support as we work together to shape the future of our community! #NLProc #Gratitude

This Thanksgiving, Iโ€™m deeply grateful for opportunity to run for NAACL Board Member alongside so many incredible candidates who share a passion for making a difference! 
I hope to earn your support as we work together to shape the future of our community! #NLProc #Gratitude
Rujun Han (@hanrujun) 's Twitter Profile Photo

Had a lot of fun working with Justin Chih-Yao Chen on reverse thinking. We show training using backward questions+reasoning with carefully designed objectives makes LLM better at a variety of reasoning tasks. Check out our paper: arxiv.org/abs/2411.19865.

Justin Chih-Yao Chen (@cyjustinchen) 's Twitter Profile Photo

Happy to share that RevThink has been accepted to #NAACL2025 main conference! ๐ŸŽ‰We also release the code and data ๐Ÿ‘‡๐Ÿงต RevThink shows that LLMs can also benefit from reverse thinking (like we often do) ๐Ÿ‘‰13.53% gains on 12 datasets (including MATH, ARC, ANLI, etc) + sample

Justin Chih-Yao Chen (@cyjustinchen) 's Twitter Profile Photo

I will be presenting โœจReverse Thinking Makes LLMs Stronger Reasonersโœจat #NAACL2025 ! We show that LLM can also benefit from reverse thinking -- a technique we often use to reason from a problem to a solution: - Improvements across 12 datasets - Outperforms SFT with 10x more

Chen-Yu Lee (@chl260) 's Twitter Profile Photo

Thrilled to introduce "๐——๐—ฒ๐—ฒ๐—ฝ ๐—ฅ๐—ฒ๐˜€๐—ฒ๐—ฎ๐—ฟ๐—ฐ๐—ต๐—ฒ๐—ฟ ๐˜„๐—ถ๐˜๐—ต ๐—ง๐—ฒ๐˜€๐˜-๐—ง๐—ถ๐—บ๐—ฒ ๐——๐—ถ๐—ณ๐—ณ๐˜‚๐˜€๐—ถ๐—ผ๐—ป," a new deep research agent designed to mimic the iterative nature of human research, complete with cycles of planning, drafting, and revision. ๐Ÿš€๐Ÿš€ arxiv.org/pdf/2507.16075

Thrilled to introduce "๐——๐—ฒ๐—ฒ๐—ฝ ๐—ฅ๐—ฒ๐˜€๐—ฒ๐—ฎ๐—ฟ๐—ฐ๐—ต๐—ฒ๐—ฟ ๐˜„๐—ถ๐˜๐—ต ๐—ง๐—ฒ๐˜€๐˜-๐—ง๐—ถ๐—บ๐—ฒ ๐——๐—ถ๐—ณ๐—ณ๐˜‚๐˜€๐—ถ๐—ผ๐—ป," a new deep research agent designed to mimic the iterative nature of human research, complete with cycles of planning, drafting, and revision. ๐Ÿš€๐Ÿš€

arxiv.org/pdf/2507.16075
Shahriar Golchin (@shahriargolchin) 's Twitter Profile Photo

Can many-shot ICL be cached and still tailored per test sample? We make it possible. ๐Ÿ’ก Excited to share that our paper, "Towards Compute-Optimal Many-Shot In-Context Learning," has been accepted to Conference on Language Modeling! Paper: arxiv.org/pdf/2507.16217 #COLM2025 #LLMs #AI #ICL

Can many-shot ICL be cached and still tailored per test sample?

We make it possible. ๐Ÿ’ก

Excited to share that our paper, "Towards Compute-Optimal Many-Shot In-Context Learning," has been accepted to <a href="/COLM_conf/">Conference on Language Modeling</a>!

Paper: arxiv.org/pdf/2507.16217

#COLM2025 #LLMs #AI #ICL
Yumo Xu (@yumo_xu) 's Twitter Profile Photo

Excited to share our #ACL2025NLP paper, "๐‚๐ข๐ญ๐ž๐„๐ฏ๐š๐ฅ: ๐๐ซ๐ข๐ง๐œ๐ข๐ฉ๐ฅ๐ž-๐ƒ๐ซ๐ข๐ฏ๐ž๐ง ๐‚๐ข๐ญ๐š๐ญ๐ข๐จ๐ง ๐„๐ฏ๐š๐ฅ๐ฎ๐š๐ญ๐ข๐จ๐ง ๐Ÿ๐จ๐ซ ๐’๐จ๐ฎ๐ซ๐œ๐ž ๐€๐ญ๐ญ๐ซ๐ข๐›๐ฎ๐ญ๐ข๐จ๐ง"! ๐Ÿ“œ If youโ€™re working on RAG, Deep Research and Trustworthy AI, this is for you. Why? Citation quality is

Excited to share our #ACL2025NLP paper, "๐‚๐ข๐ญ๐ž๐„๐ฏ๐š๐ฅ: ๐๐ซ๐ข๐ง๐œ๐ข๐ฉ๐ฅ๐ž-๐ƒ๐ซ๐ข๐ฏ๐ž๐ง ๐‚๐ข๐ญ๐š๐ญ๐ข๐จ๐ง ๐„๐ฏ๐š๐ฅ๐ฎ๐š๐ญ๐ข๐จ๐ง ๐Ÿ๐จ๐ซ ๐’๐จ๐ฎ๐ซ๐œ๐ž ๐€๐ญ๐ญ๐ซ๐ข๐›๐ฎ๐ญ๐ข๐จ๐ง"! ๐Ÿ“œ If youโ€™re working on RAG, Deep Research and Trustworthy AI, this is for you. Why? Citation quality is