hyunji amy lee (@hyunji_amy_lee) 's Twitter Profile
hyunji amy lee

@hyunji_amy_lee

PhD student @kaist_ai. Previously: @allen_ai @Adobe.

ID: 1407986466018398209

linkhttps://amy-hyunji.github.io/ calendar_today24-06-2021 08:58:38

113 Tweet

629 Followers

399 Following

hyunji amy lee (@hyunji_amy_lee) 's Twitter Profile Photo

Wonder why DPO works so well? Check out our paper led by Yunjae Won for deep insights into its effectiveness and behavior from an information-theoretic perspective!

Dayeon (Zoey) Ki (@zoeykii) 's Twitter Profile Photo

1/ Are two #LLMs better than one for equitable cultural alignment? 🌍 We introduce a Multi-Agent Debate framework β€” where two LLM agents debate the cultural adaptability of a given scenario. #ACL2025 πŸ§΅πŸ‘‡

1/ Are two #LLMs better than one for equitable cultural alignment? 🌍 

We introduce a Multi-Agent Debate framework β€” where two LLM agents debate the cultural adaptability of a given scenario.

#ACL2025 πŸ§΅πŸ‘‡
Sohee Yang (@soheeyang_) 's Twitter Profile Photo

🚨 New Paper 🧡 How effectively do reasoning models reevaluate their thought? We find that: - Models excel at identifying unhelpful thoughts but struggle to recover from them - Smaller models can be more robust - Self-reevaluation ability is far from true meta-cognitive awareness

🚨 New Paper 🧡
How effectively do reasoning models reevaluate their thought? We find that:
- Models excel at identifying unhelpful thoughts but struggle to recover from them
- Smaller models can be more robust
- Self-reevaluation ability is far from true meta-cognitive awareness
Lucas Caccia (@lucaspcaccia) 's Twitter Profile Photo

RAG and in-context learning are the go-to approaches for integrating new knowledge into LLMs, making inference very inefficient We propose instead π—žπ—»π—Όπ˜„π—Ήπ—²π—±π—΄π—² π— π—Όπ—±π˜‚π—Ήπ—²π˜€ : lightweight LoRA modules trained offline that can match RAG performance without the drawbacks

Mohit Bansal (@mohitban47) 's Twitter Profile Photo

πŸŽ‰ Yay, welcome hyunji amy lee -- super excited to have you join us as a postdoc! πŸ€— Welcome to our MURGe-Lab + UNC AI + UNC Computer Science family & the beautiful Research Triangle area -- looking forward to the many fun+impactful collaborations together πŸ”₯

Mohit Bansal (@mohitban47) 's Twitter Profile Photo

PS. FYI, Hyunji (Amy)'s expertise/interests are in retrieval, aligning knowledge modules with LLM's parametric knowledge, and expanding to various modalities, with diverse+extensive work at KAIST, AI2, Adobe, MSR, etc., details at --> amy-hyunji.github.io

Duy Nguyen (@duynguyen772) 's Twitter Profile Photo

πŸš€ We introduce GrAInS, a gradient-based attribution method for inference-time steering (of both LLMs & VLMs). βœ… Works for both LLMs (+13.2% on TruthfulQA) & VLMs (+8.1% win rate on SPA-VL). βœ… Preserves core abilities (<1% drop on MMLU/MMMU). LLMs & VLMs often fail because

πŸš€ We introduce GrAInS, a gradient-based attribution method for inference-time steering (of both LLMs &amp; VLMs).

βœ… Works for both LLMs (+13.2% on TruthfulQA) &amp; VLMs (+8.1% win rate on SPA-VL).
βœ… Preserves core abilities (&lt;1% drop on MMLU/MMMU).

LLMs &amp; VLMs often fail because
Vishakh Padmakumar (@vishakh_pk) 's Twitter Profile Photo

Maybe don't use an LLM for _everything_? Last summer, I got to fiddle again with content diversity Adobe Research Adobe and we showed that agentic pipelines that mix LLM-prompt steps with principled techniques can yield better, more personalized summaries

Maybe don't use an LLM for _everything_?

Last summer, I got to fiddle again with content diversity <a href="/AdobeResearch/">Adobe Research</a> <a href="/Adobe/">Adobe</a> and we showed that agentic pipelines that mix LLM-prompt steps with principled techniques can yield better, more personalized summaries