Jonathan Lai (@_jlai) 's Twitter Profile
Jonathan Lai

@_jlai

Post training @GoogleDeepMind, Gemini Reasoning, RL, Opinions are my own

ID: 971441900

calendar_today26-11-2012 06:31:19

18 Tweet

115 Followers

138 Following

Jonathan Lai (@_jlai) 's Twitter Profile Photo

🚨✨ Thrilled to share the first study on model merging at large scales by our intern Prateek Yadav Google AI Google DeepMind For larger LLMs merging is an efficient alternative to multitask learning, that can preserve the majority of in-domain performance, while significantly

Prateek Yadav (@prateeky2806) 's Twitter Profile Photo

I'm on the job market! Please reach out if you are looking to hire someone to work on - RLHF - Efficiency - MoE/Modular models - Synthetic Data - Test time compute - other phases of pre/post-training. If you are not hiring then I would appreciate a retweet! More details👇

Logan Kilpatrick (@officiallogank) 's Twitter Profile Photo

We are rolling out a new Gemini 2.0 Flash Thinking update: - Exp-01-21 variant in AI Studio and API for free - 1 million token context window - Native code execution support - Longer output token generation - Less frequent model contradictions Try it aistudio.google.com

Sundar Pichai (@sundarpichai) 's Twitter Profile Photo

1/ Gemini 2.5 is here, and it’s our most intelligent AI model ever. Our first 2.5 model, Gemini 2.5 Pro Experimental is a state-of-the-art thinking model, leading in a wide range of benchmarks – with impressive improvements in enhanced reasoning and coding and now #1 on

Tsendsuren (@tsendeemts) 's Twitter Profile Photo

Gemini-ийн шинэ загварыг туршаад үзээрэй. Код бичих дээр нилээн сайжирсан байгаа.

Tu Vu (@tuvllms) 's Twitter Profile Photo

🚨 New paper 🚨 Excited to share my first paper w/ my PhD students!! We find that advanced LLM capabilities conferred by instruction or alignment tuning (e.g., SFT, RLHF, DPO, GRPO) can be encoded into model diff vectors (à la task vectors) and transferred across model

🚨 New paper 🚨

Excited to share my first paper w/ my PhD students!!

We find that advanced LLM capabilities conferred by instruction or alignment tuning (e.g., SFT, RLHF, DPO, GRPO) can be encoded into model diff vectors (à la task vectors) and transferred across model
Tu Vu (@tuvllms) 's Twitter Profile Photo

Excited to share that our paper on model merging at scale has been accepted to Transactions on Machine Learning Research (TMLR). Huge congrats to my intern Prateek Yadav and our awesome co-authors Jonathan Lai, Alexandra Chronopoulou, Manaal Faruqui, Mohit Bansal, and Tsendsuren 🎉!!

Excited to share that our paper on model merging at scale has been accepted to Transactions on Machine Learning Research (TMLR). Huge congrats to my intern <a href="/prateeky2806/">Prateek Yadav</a> and our awesome co-authors <a href="/_JLai/">Jonathan Lai</a>, <a href="/alexandraxron/">Alexandra Chronopoulou</a>, <a href="/manaalfar/">Manaal Faruqui</a>, <a href="/mohitban47/">Mohit Bansal</a>, and <a href="/TsendeeMTS/">Tsendsuren</a> 🎉!!