Seongyun Lee (@sylee_ai) 's Twitter Profile
Seongyun Lee

@sylee_ai

Research intern @LG_AI_Research | M.S. Student @kaist_ai | Researching on reasoning & planning abilities of LLM/VLM #NLProc

ID: 1695420388384366594

calendar_today26-08-2023 12:58:29

466 Tweet

261 Followers

305 Following

Seungone Kim @ NAACL2025 (@seungonekim) 's Twitter Profile Photo

#NLProc New paper on "evaluation-time scaling", a new dimension to leverage test-time compute! We replicate the test-time scaling behaviors observed in generators (e.g., o1, r1, s1) with evaluators by enforcing to generate additional reasoning tokens. arxiv.org/abs/2503.19877

#NLProc 
New paper on "evaluation-time scaling", a new dimension to leverage test-time compute!

We replicate the test-time scaling behaviors observed in generators (e.g., o1, r1, s1) with evaluators by enforcing to generate additional reasoning tokens.

arxiv.org/abs/2503.19877
Geewook Kim (@geewookkim) 's Twitter Profile Photo

Presenting our poster at #ICLR2025 today (Fri, Apr 25, 15:00) — Hall 3 + Hall 2B #264! We explored safety issues when extending LLMs to vision and how to address them. Come by and let’s chat—always happy to discuss ideas! 🤗

Presenting our poster at #ICLR2025 today (Fri, Apr 25, 15:00) — Hall 3 + Hall 2B #264! 

We explored safety issues when extending LLMs to vision and how to address them. Come by and let’s chat—always happy to discuss ideas! 🤗
Jinheon Baek (@jinheonbaek) 's Twitter Profile Photo

So excited to drop PaperCoder, a multi-agent LLM system that turns ML papers into full codebases. It looks like this:📄 (papers) → 🧠 (planning) → 🛠️ (full repos), all powered by 🤖. Big thanks to AK for the shoutout! Paper: arxiv.org/abs/2504.17192

Seungone Kim @ NAACL2025 (@seungonekim) 's Twitter Profile Photo

🏆Glad to share that our BiGGen Bench paper has received the best paper award at NAACL HLT 2025! x.com/naaclmeeting/s… 📅 Ballroom A, Session I: Thursday May 1st, 16:00-17:30 (MDT) 📅 Session M (Plenary Session): Friday May 2nd, 15:30-16:30 (MDT) 📅 Virtual Conference: Tuesday

jiyeon kim (@jiyeonkimd) 's Twitter Profile Photo

Presenting ✨Knowledg Entropy✨ at #ICLR2025 today in Oral 5C(Garnet 216-218) at 10:30AM and in Poster 6(#251) from 3:00PM We investigated how changes in a model's tendency to integrate its parametric knowledge during pretraining affect knowledge acquisition and forgetting

Presenting ✨Knowledg Entropy✨ at #ICLR2025 today in Oral 5C(Garnet 216-218) at 10:30AM and in Poster 6(#251) from 3:00PM

We investigated how changes in a model's tendency to integrate its parametric knowledge during pretraining affect knowledge acquisition and forgetting
elvis (@omarsar0) 's Twitter Profile Photo

The CoT Encyclopedia How to predict and steer the reasoning strategies of LLMs that use chain-of-thought (CoT)? More below:

The CoT Encyclopedia

How to predict and steer the reasoning strategies of LLMs that use chain-of-thought (CoT)?

More below:
fly51fly (@fly51fly) 's Twitter Profile Photo

[CL] The CoT Encyclopedia: Analyzing, Predicting, and Controlling how a Reasoning Model will Think S Lee, S Kim, M Seo, Y Jo... [KAIST AI & CMU] (2025) arxiv.org/abs/2505.10185

[CL] The CoT Encyclopedia: Analyzing, Predicting, and Controlling how a Reasoning Model will Think
S Lee, S Kim, M Seo, Y Jo... [KAIST AI & CMU] (2025)
arxiv.org/abs/2505.10185
The AI Timeline (@theaitimeline) 's Twitter Profile Photo

🚨This week's top AI/ML research papers: - AlphaEvolve - Qwen3 Technical Report - Insights into DeepSeek-V3 - Seed1.5-VL Technical Report - BLIP3-o - Parallel Scaling Law for LMs - HealthBench - Learning Dynamics in Continual Pre-Training for LLMs - Learning to Think - Beyond

🚨This week's top AI/ML research papers:

- AlphaEvolve
- Qwen3 Technical Report
- Insights into DeepSeek-V3
- Seed1.5-VL Technical Report
- BLIP3-o
- Parallel Scaling Law for LMs
- HealthBench
- Learning Dynamics in Continual Pre-Training for LLMs
- Learning to Think
- Beyond
Minki Kang (@mkkang_1133) 's Twitter Profile Photo

🚨 New preprint! Can small language models (sLMs) solve complex problems like LLMs? We show how to go beyond cloning reasoning—to distill tool-using agent behavior into sLMs as tiny as 0.5B. Meet Agent Distillation: 📄 huggingface.co/papers/2505.17… Here's the details 🧵👇:

Changdae Oh (@changdae_oh) 's Twitter Profile Photo

Does anyone want to dig deeper into the robustness of Multimodal LLMs (MLLMs) beyond empirical observations Happy to serve this exactly through our new #ICML2025 paper "Understanding Multimodal LLMs Under Distribution Shifts: An Information-Theoretic Approach"!

Does anyone want to dig deeper into the robustness of Multimodal LLMs (MLLMs) beyond empirical observations

Happy to serve this exactly through our new #ICML2025 paper "Understanding Multimodal LLMs Under Distribution Shifts: An Information-Theoretic Approach"!
Hyeonbin Hwang (@ronalhwang) 's Twitter Profile Photo

🚨 New Paper co-led with byeongguk jeon 🚨 Q. Can we adapt Language Models, trained to predict next token, to reason in sentence-level? I think LMs operating in higher-level abstraction would be a promising path towards advancing its reasoning, and I am excited to share our

🚨 New Paper co-led with <a href="/bkjeon1211/">byeongguk jeon</a> 🚨

Q. Can we adapt Language Models, trained to predict next token, to reason in sentence-level? 

I think LMs operating in higher-level abstraction would be a promising path towards advancing its reasoning, and I am excited to share our
hyunji amy lee (@hyunji_amy_lee) 's Twitter Profile Photo

🚨 Want models to better utilize and ground on the provided knowledge? We introduce Context-INformed Grounding Supervision (CINGS)! Training LLM with CINGS significantly boosts grounding abilities in both text and vision-language models compared to standard instruction tuning.

🚨 Want models to better utilize and ground on the provided knowledge? We introduce Context-INformed Grounding Supervision (CINGS)! Training LLM with CINGS significantly boosts grounding abilities in both text and vision-language models compared to standard instruction tuning.
Sakana AI (@sakanaailabs) 's Twitter Profile Photo

We’re excited to introduce AB-MCTS! Our new inference-time scaling algorithm enables collective intelligence for AI by allowing multiple frontier models (like Gemini 2.5 Pro, o4-mini, DeepSeek-R1-0528) to cooperate. Blog: sakana.ai/ab-mcts Paper: arxiv.org/abs/2503.04412

We’re excited to introduce AB-MCTS!

Our new inference-time scaling algorithm enables collective intelligence for AI by allowing multiple frontier models (like Gemini 2.5 Pro, o4-mini, DeepSeek-R1-0528) to cooperate.

Blog: sakana.ai/ab-mcts
Paper: arxiv.org/abs/2503.04412