Juno KIM (@junokim_ai) 's Twitter Profile
Juno KIM

@junokim_ai

Incoming EECS PhD student @UCBerkeley research: deep learning theory

ID: 1751494597870858240

linkhttps://junokim1.github.io/ calendar_today28-01-2024 06:37:24

42 Tweet

218 Followers

102 Following

Juno KIM (@junokim_ai) 's Twitter Profile Photo

Also giving a contributed talk on the learning-theoretic complexity and optimality of ICL at the Theoretical Foundations of Foundation Models happy to share our results with the ML theory and LLM community!

Also giving a contributed talk on the learning-theoretic complexity and optimality of ICL at the <a href="/tf2m_workshop/">Theoretical Foundations of Foundation Models</a> happy to share our results with the ML theory and LLM community!
Zeyuan Allen-Zhu, Sc.D. (@zeyuanallenzhu) 's Twitter Profile Photo

Incredibly honored and humbled by the overwhelming response to my tutorial, and thank you everyone who attended in person. Truly heartwarming to hear how much you enjoyed it. Many have been asking for a recording, and I prepared one with my own subtitles youtu.be/yBL7J0kgldU

Incredibly honored and humbled by the overwhelming response to my tutorial, and thank you everyone who attended in person. Truly heartwarming to hear how much you enjoyed it. Many have been asking for a recording, and I prepared one with my own subtitles  youtu.be/yBL7J0kgldU
Alex Bilzerian (@alexbilz) 's Twitter Profile Photo

'High-Dimensional Probability' - Vershynin (2024, PDF): math.uci.edu/~rvershyn/pape… Full course on high-dimensional probability with 41 video lectures & 13 problem sets provided: math.uci.edu/~rvershyn/teac…

'High-Dimensional Probability' - Vershynin (2024, PDF): math.uci.edu/~rvershyn/pape…

Full course on high-dimensional probability with 41 video lectures &amp; 13 problem sets provided: math.uci.edu/~rvershyn/teac…
Juno KIM (@junokim_ai) 's Twitter Profile Photo

Our paper on statistical complexity and optimality of in-context learning of deep transformers has been accepted to NeurIPS! arxiv.org/abs/2408.12186

Taiji Suzuki (@btreetaiji) 's Twitter Profile Photo

Here is a great Youtube video by Charles Riou that effectively explains our recent paper with Juno KIM on chain-of-thought accepted by ICLR2025. Check it! Kim&Suzuki: Transformers Provably Solve Parity Efficiently with Chain of Thought. ICLR2025. youtu.be/pj-GEBU2iVs?si…

Sasha Rush (@srush_nlp) 's Twitter Profile Photo

Simons Institute Workshop: "Future of LLMs and Transformers": 21 talks Monday - Friday next week. simons.berkeley.edu/workshops/futu…

Simons Institute Workshop: "Future of LLMs and Transformers": 21 talks Monday - Friday next week.

simons.berkeley.edu/workshops/futu…
Juno KIM (@junokim_ai) 's Twitter Profile Photo

Graduated with my master's from the University of Tokyo!🇯🇵 I was honored to receive the graduate school dean's award for outstanding research and to deliver a speech at the ceremony! Deeply grateful to my supervisor Taiji Suzuki🙏 i.u-tokyo.ac.jp/news/topics/20…

Juno KIM (@junokim_ai) 's Twitter Profile Photo

Our paper studying dynamics of a simple Markov model for CoT reasoning has been accepted to #ICML2025 ! (reposting a nice summary↓)

Zixuan Wang (@zzzixuanwang) 's Twitter Profile Photo

LLMs can solve complex tasks that require combining multiple reasoning steps. But when are such capabilities learnable via gradient-based training? In our new COLT 2025 paper, we show that easy-to-hard data is necessary and sufficient! arxiv.org/abs/2505.23683 🧵 below (1/10)

LLMs can solve complex tasks that require combining multiple reasoning steps. But when are such capabilities learnable via gradient-based training?

In our new COLT 2025 paper, we show that easy-to-hard data is necessary and sufficient!

arxiv.org/abs/2505.23683

🧵 below (1/10)
Jyo Pari (@jyo_pari) 's Twitter Profile Photo

What if an LLM could update its own weights? Meet SEAL🦭: a framework where LLMs generate their own training data (self-edits) to update their weights in response to new inputs. Self-editing is learned via RL, using the updated model’s downstream performance as reward.

What if an LLM could update its own weights?

Meet SEAL🦭: a framework where LLMs generate their own training data (self-edits) to update their weights in response to new inputs.

Self-editing is learned via RL, using the updated model’s downstream performance as reward.