Dongkeun Yoon (@dongkeun_yoon) 's Twitter Profile
Dongkeun Yoon

@dongkeun_yoon

PhD student @kaist_ai, research intern @LG_AI_Research. Researching multilinguality in LLMs.

ID: 1504088658135040008

linkhttps://mattyoon.github.io/ calendar_today16-03-2022 13:34:20

127 Tweet

324 Followers

239 Following

Dongkeun Yoon (@dongkeun_yoon) 's Twitter Profile Photo

🙁 LLMs are overconfident even when they are dead wrong. 🧐 What about reasoning models? Can they actually tell us “My answer is only 60% likely to be correct”? ❗Our paper suggests that they can! Through extensive analysis, we investigate what enables this emergent ability.

🙁 LLMs are overconfident even when they are dead wrong.

🧐 What about reasoning models? Can they actually tell us “My answer is only 60% likely to be correct”?

❗Our paper suggests that they can! Through extensive analysis, we investigate what enables this emergent ability.
Sohee Yang (@soheeyang_) 's Twitter Profile Photo

Reasoning models are quite verbose in their thinking process. Is it any good? We find out that it enables reasoning models to be more accurate in telling what they know and don’t know (confidence)! Even non-reasoning models can do it better if they mimic the verbose reasoning! 👀

Seungone Kim @ NAACL2025 (@seungonekim) 's Twitter Profile Photo

Turns out that reasoning models not only excel at solving problems but are also excellent confidence estimators - an unexpected side effect of long CoTs! This reminds me that smart ppl are good at determining what they know & don't know👀 Check out Dongkeun Yoon 's post!

fly51fly (@fly51fly) 's Twitter Profile Photo

[CL] Reasoning Models Better Express Their Confidence D Yoon, S Kim, S Yang, S Kim... [KAIST & CMU & UCL] (2025) arxiv.org/abs/2505.14489

[CL] Reasoning Models Better Express Their Confidence
D Yoon, S Kim, S Yang, S Kim... [KAIST & CMU & UCL] (2025)
arxiv.org/abs/2505.14489
Smells Like ML (@smellslikeml) 's Twitter Profile Photo

Dongkeun Yoon Congrats to the team for this fantastic work! Had a chance to try the code on my reasoning VLM and found consistent results. x.com/smellslikeml/s…

Chaeeun Kim (@chaechaek1214) 's Twitter Profile Photo

❓What if your RAG didn’t need a separate retrieval model at all? We present 🧊FREESON, a new framework for retriever-FREE retrieval-augmented reasoning. With FREESON, a single LRM acts as both generator and retriever, shifting the focus from seq2seq matching to locating

❓What if your RAG didn’t need a separate retrieval model at all?

We present 🧊FREESON, a new framework for retriever-FREE retrieval-augmented reasoning.

With FREESON,  a single LRM acts as both generator and retriever, shifting the focus from seq2seq matching to locating
arlo_son (@gson_ai) 's Twitter Profile Photo

Imagine you’re collaborating with an AI co-scientist: you ask it to proofread your manuscript and flag any errors. Which LLM would you choose? 🤔 We evaluated the new Claude 4 models on SPOT. It looks like o3 is still the best model for this.

Imagine you’re collaborating with an AI co-scientist: you ask it to proofread your manuscript and flag any errors. Which LLM would you choose? 🤔

We evaluated the new Claude 4 models on SPOT. It looks like o3 is still the best model for this.
Hoyeon Chang (@hoyeon_chang) 's Twitter Profile Photo

New preprint 📄 (with Jinho Park ) Can neural nets really reason compositionally, or just match patterns? We present the Coverage Principle: a data-centric framework that predicts when pattern-matching models will generalize (validated on Transformers). 🧵👇

New preprint 📄 (with <a href="/jinho___park/">Jinho Park</a> )

Can neural nets really reason compositionally, or just match patterns?  
We present the Coverage Principle: a data-centric framework that predicts when pattern-matching models will generalize (validated on Transformers). 🧵👇
Hyeonbin Hwang (@ronalhwang) 's Twitter Profile Photo

🚨 New Paper co-led with byeongguk jeon 🚨 Q. Can we adapt Language Models, trained to predict next token, to reason in sentence-level? I think LMs operating in higher-level abstraction would be a promising path towards advancing its reasoning, and I am excited to share our

🚨 New Paper co-led with <a href="/bkjeon1211/">byeongguk jeon</a> 🚨

Q. Can we adapt Language Models, trained to predict next token, to reason in sentence-level? 

I think LMs operating in higher-level abstraction would be a promising path towards advancing its reasoning, and I am excited to share our
Sheikh Shafayat ✈️ ICLR'25 🇸🇬 (@shafayat_sheikh) 's Twitter Profile Photo

Check out our latest work on self-improving LLMs, where we try to see if LLMs can utilize their internal self consistency as a reward signal to bootstrap itself using RL. TL;DR: it can, to some extent, but then ends up reward hacking the self-consistency objective. We try to see

Check out our latest work on self-improving LLMs, where we try to see if LLMs can utilize their internal self consistency as a reward signal to bootstrap itself using RL.

TL;DR: it can, to some extent, but then ends up reward hacking the self-consistency objective. We try to see
Dayoon Ko (@dayoon12161) 's Twitter Profile Photo

🚨 Excited to share that our paper was accepted to #ACL2025 Findings 🎉 "When Should Dense Retrievers Be Updated in Evolving Corpora? Detecting Out-of-Distribution Corpora Using GradNormIR" Huge thanks to my amazing collaborators! 🙌 Jinyoung Kim Sohyeon Kim We propose

Sohee Yang (@soheeyang_) 's Twitter Profile Photo

🚨 New Paper 🧵 How effectively do reasoning models reevaluate their thought? We find that: - Models excel at identifying unhelpful thoughts but struggle to recover from them - Smaller models can be more robust - Self-reevaluation ability is far from true meta-cognitive awareness

🚨 New Paper 🧵
How effectively do reasoning models reevaluate their thought? We find that:
- Models excel at identifying unhelpful thoughts but struggle to recover from them
- Smaller models can be more robust
- Self-reevaluation ability is far from true meta-cognitive awareness
hyunji amy lee (@hyunji_amy_lee) 's Twitter Profile Photo

🚨 Want models to better utilize and ground on the provided knowledge? We introduce Context-INformed Grounding Supervision (CINGS)! Training LLM with CINGS significantly boosts grounding abilities in both text and vision-language models compared to standard instruction tuning.

🚨 Want models to better utilize and ground on the provided knowledge? We introduce Context-INformed Grounding Supervision (CINGS)! Training LLM with CINGS significantly boosts grounding abilities in both text and vision-language models compared to standard instruction tuning.
Ricardo Rei (@ricardorei7) 's Twitter Profile Photo

🚀 Tower+: our latest model in the Tower family — sets a new standard for open-weight multilingual models! We show how to go beyond sentence-level translation, striking a balance between translation quality and general multilingual capabilities. 1/5 arxiv.org/pdf/2506.17080

🚀 Tower+: our latest model in the Tower family — sets a new standard for open-weight multilingual models!
We show how to go beyond sentence-level translation, striking a balance between translation quality and general multilingual capabilities.
1/5

arxiv.org/pdf/2506.17080
José Maria Pombal (@zmprcp) 's Twitter Profile Photo

Check out the latest iteration of Tower models, Tower+. Ideal for translation tasks and beyond, and available at three different scales: 2B, 9B, 72B. All available on huggingface: huggingface.co/collections/Un… Kudos to everyone involved!

hyunji amy lee (@hyunji_amy_lee) 's Twitter Profile Photo

🥳Excited to share that I’ll be joining UNC Computer Science as postdoc this fall. Looking forward to work with Mohit Bansal & amazing students at UNC AI. I'll continue working on retrieval, aligning knowledge modules with LLM's parametric knowledge, and expanding to various modalities.

🥳Excited to share that I’ll be joining <a href="/unccs/">UNC Computer Science</a> as postdoc this fall. Looking forward to work with <a href="/mohitban47/">Mohit Bansal</a> &amp; amazing students at <a href="/unc_ai_group/">UNC AI</a>.
I'll continue working on retrieval, aligning knowledge modules with LLM's parametric knowledge, and expanding to various modalities.