Seong Joon Oh (@coallaoh) 's Twitter Profile
Seong Joon Oh

@coallaoh

Professor in Scalable Trustworthy AI @ University of Tübingen | Advisor at Parameter Lab & ResearchTrend.AI

ID: 760050275289862148

linkhttp://seongjoonoh.com calendar_today01-08-2016 09:51:21

1,1K Tweet

2,2K Followers

1,1K Following

Namhoon Lee (@namhoonlee09) 's Twitter Profile Photo

🔥NEW ICLR 2025 PAPER! ZIP: An Efficient Zeroth-order Prompt Tuning for Black-box Vision-Language Models paper: arxiv.org/abs/2504.06838 poster: #377, Poster session 2, Thursday 3-5:30pm, Hall 3 + Hall 2B thread 🧵

Seong Joon Oh (@coallaoh) 's Twitter Profile Photo

The era of generalist AI paves the way to private AI and adaptation. Generalist AI solves IMO problems better than any human. Yet on my email threads it produces rubbish. I cannot fully delegate my work pipeline to AI. This happens because of adaptation issues; we must bring

Alexander Rubinstein (@a_rubique) 's Twitter Profile Photo

Glad to be at #ICLR2025! Ankit Sonthalia and I are presenting our paper on star-shapeness of deep learning solution sets today at poster session 5 (poster #132)! Feel free to drop by and talk about the structure of the models' weight space, generalization and uncertainty!

Haritz Puerto @ NAACL 2025 🌵🇺🇸 (@haritzpuerto) 's Twitter Profile Photo

I will be in person at #NAACL2025 🌵🇺🇸 to present Scaling Up Membership Inference: When and How Attacks Succeed on LLMs. Come and say hi 👋 if you want to know how to proof if an LLM was trained on a data point!

Bhaktavaschal Samal (@bhaktavee) 's Twitter Profile Photo

here is my take after skimming through the 18 page article- collection-level mia can work even though gpt-style models that see only 4 k–200 k tokens at once. what to need basically is to chunk a ~1 million token corpus, score each chunk (loss, confidence, embedding drift) and

Seong Joon Oh (@coallaoh) 's Twitter Profile Photo

🔴 LIVE NOW! researchtrend.ai/zoom 🤖 LLMAG Session (16:00 - 17:00 UTC) 🖼️ VLM Session (17:00 - 18:00 UTC) 🎯 FedML Session (18:00 - 19:00 UTC)

Kevin Patrick Murphy (@sirbayes) 's Twitter Profile Photo

I am pleased to announce a new version of my RL tutorial. Major update to the LLM chapter (eg DPO, GRPO, thinking), minor updates to the MARL and MBRL chapters and various sections (eg offline RL, DPG, etc). Enjoy! arxiv.org/abs/2412.05265

I am pleased to announce a new version of my RL tutorial. Major update to the LLM chapter (eg DPO, GRPO, thinking), minor updates to the MARL and MBRL chapters and various sections (eg offline RL, DPG, etc). Enjoy!
arxiv.org/abs/2412.05265
Sagnik Mukherjee (@saagnikkk) 's Twitter Profile Photo

🚨 Paper Alert: “RL Finetunes Small Subnetworks in Large Language Models” From DeepSeek V3 Base to DeepSeek R1 Zero, a whopping 86% of parameters were NOT updated during RL training 😮😮 And this isn’t a one-off. The pattern holds across RL algorithms and models. 🧵A Deep Dive

🚨 Paper Alert: “RL Finetunes Small Subnetworks in Large Language Models”

From DeepSeek V3 Base to DeepSeek R1 Zero, a whopping 86% of parameters were NOT updated during RL training 😮😮
And this isn’t a one-off. The pattern holds across RL algorithms and models.
🧵A Deep Dive
Seong Joon Oh (@coallaoh) 's Twitter Profile Photo

After 6 NeurIPS submissions, 5 ICCV rebuttals, and 1 EMNLP submission, I’m trying to regain sanity. Handling 400+ emails - Done. Now, let's sort out those 169 todos. AI says the estimated average duration for each item is 2.9 hours.

Sohee Yang (@soheeyang_) 's Twitter Profile Photo

Reasoning models are quite verbose in their thinking process. Is it any good? We find out that it enables reasoning models to be more accurate in telling what they know and don’t know (confidence)! Even non-reasoning models can do it better if they mimic the verbose reasoning! 👀

Seong Joon Oh (@coallaoh) 's Twitter Profile Photo

Congrats 🎉 Happy to announce a paper with Yujin, Arnas, and Anna.  It was such a fun collaboration. Our paper tackles a fundamental question raised by the Generative AI Paradox (ICLR 2024, arxiv.org/abs/2311.00059): “What it can create, can it also understand?” Some recent

Seong Joon Oh (@coallaoh) 's Twitter Profile Photo

Recently updated my website seongjoonoh.com using community tags from ResearchTrend.AI. It reflects the evolution of my research interests and highlights my current focus. Visitors can browse topics of interest with ease. At ResearchTrend.AI, we're

Seong Joon Oh (@coallaoh) 's Twitter Profile Photo

🔴 LIVE NOW at researchtrend.ai/zoom We've got great speakers: Yuntao Du from Purdue University Shaina Raza, PhD from Vector Institute David Fan from Meta Youssef Allouah from EPFL We cover fresh papers on LLM Agents, VLM, & Federated Learning. Think of it as a giant reading

Kyunghyun Cho (@kchonyc) 's Twitter Profile Photo

the new administration in Korea has managed to bring onboard Jung-Woo Ha as the secretary of AI and future technologies to directly report to and advise the President. some real AI researcher, developer and director! a great job!

the new administration in Korea has managed to bring onboard <a href="/JungWooHa2/">Jung-Woo Ha</a> as the secretary of AI and future technologies to directly report to and advise the President. some real AI researcher, developer and director! a great job!
Seong Joon Oh (@coallaoh) 's Twitter Profile Photo

🔴 LIVE NOW at researchtrend.ai/zoom Event agenda: researchtrend.ai/social-events/… ResearchTrend.AI #LLM #VLM #FederatedLearning #AI

🔴 LIVE NOW at researchtrend.ai/zoom 

Event agenda: researchtrend.ai/social-events/…

ResearchTrend.AI

#LLM #VLM #FederatedLearning #AI