Irena Gao (@irena_gao) 's Twitter Profile
Irena Gao

@irena_gao

PhD student @StanfordAILab | Trustworthy ML

ID: 1573342259604357121

linkhttp://i-gao.github.io calendar_today23-09-2022 16:03:30

99 Tweet

972 Followers

254 Following

Luke Bailey (@lukebailey181) 's Twitter Profile Photo

Understanding the current landscape of AI agents is important for predicting future trends in this area. Read our paper if you want a snapshot of deployed AI agents.

Stanford AI Lab (@stanfordailab) 's Twitter Profile Photo

SAIL is delighted to announce Carlos Carlos Guestrin, the Fortinet Founders Professor of Computer Science, as the next Director of Stanford AI Lab. Carlos is a talented researcher and leader, known for his work on explainability, graphs, compilation, and boosted trees in AI.

SAIL is delighted to announce Carlos <a href="/guestrin/">Carlos Guestrin</a>, the Fortinet Founders Professor of Computer Science, as the next Director of <a href="/StanfordAILab/">Stanford AI Lab</a>. Carlos is a talented researcher and leader, known for his work on explainability, graphs, compilation, and boosted trees in AI.
CLS (@chengleisi) 's Twitter Profile Photo

Reasoning is all the rage these days. If you want to save some time and get to the crux of how to enable reasoning in LLMs, here’s a list of 10 recent papers that I find most informative, along with my notes: (Full thread in doc: docs.google.com/document/d/1TW…) 1/11

Anikait Singh (@anikait_singh_) 's Twitter Profile Photo

Personalization in LLMs is crucial for meeting diverse user needs, yet collecting real-world preferences at scale remains a significant challenge. Introducing FSPO, a simple framework leveraging synthetic preference data to adapt new users with meta-learning for open-ended QA! 🧵

Personalization in LLMs is crucial for meeting diverse user needs, yet collecting real-world preferences at scale remains a significant challenge. Introducing FSPO, a simple framework leveraging synthetic preference data to adapt new users with meta-learning for open-ended QA! 🧵
Eddie Vendrow (@edwardvendrow) 's Twitter Profile Photo

Very excited to share *GSM8K-Platinum*, a revised version of the GSM8K test set! If you’re using GSM8K, I highly recommend you switch to GSM8K-Platinum! We built it as a drop-in replacement for the GSM8K test set. Check it out: huggingface.co/datasets/madry…

James Zou (@james_y_zou) 's Twitter Profile Photo

💡The key idea of #textgrad is to optimize by backpropagating textual gradients produced by #LLM. Paper: nature.com/articles/s4158… Code: github.com/zou-group/text… Amazing job by Mert Yuksekgonul leading this project w/ fantastic collaborators Federico Bianchi Joseph Boen Sheng Liu

Stanford AI Lab (@stanfordailab) 's Twitter Profile Photo

Heading to #ICLR2025 ? Make sure to check out the amazing research led by our students here at the Stanford AI Lab! ai.stanford.edu/blog/iclr-2025/

Amber Xie (@amberxie_) 's Twitter Profile Photo

Introducing ✨Latent Diffusion Planning✨ (LDP)! We explore how to use expert, suboptimal, & action-free data. To do so, we learn a diffusion-based *planner* that forecasts latent states, and an *inverse-dynamics model* that extracts actions. w/ Oleg Rybkin Dorsa Sadigh Chelsea Finn

Fahim Tajwar (@fahimtajwar10) 's Twitter Profile Photo

RL with verifiable reward has shown impressive results in improving LLM reasoning, but what can we do when we do not have ground truth answers? Introducing Self-Rewarding Training (SRT): where language models provide their own reward for RL training! 🧵 1/n

RL with verifiable reward has shown impressive results in improving LLM reasoning, but what can we do when we do not have ground truth answers?

Introducing Self-Rewarding Training (SRT): where language models provide their own reward for RL training!

🧵 1/n
Qinan Yu (@qinan_yu) 's Twitter Profile Photo

🎀 fine-grained, interpretable representation steering for LMs! meet RePS — Reference-free Preference Steering! 1⃣ outperforms existing methods on 2B-27B LMs, nearly matching prompting 2⃣ supports both steering and suppression (beat system prompts!) 3⃣ jailbreak-proof (1/n)

🎀 fine-grained, interpretable representation steering for LMs!
meet RePS — Reference-free Preference Steering!

1⃣ outperforms existing methods on 2B-27B LMs, nearly matching prompting
2⃣ supports both steering and suppression (beat system prompts!)
3⃣ jailbreak-proof

(1/n)
Omar Shaikh (@oshaikh13) 's Twitter Profile Photo

What if LLMs could learn your habits and preferences well enough (across any context!) to anticipate your needs? In a new paper, we present the General User Model (GUM): a model of you built from just your everyday computer use. 🧵

Yijia Shao (@echoshao8899) 's Twitter Profile Photo

🚨 70 million US workers are about to face their biggest workplace transmission due to AI agents. But nobody asks them what they want. While AI races to automate everything, we took a different approach: auditing what workers want vs. what AI can do across the US workforce.🧵

🚨 70 million US workers are about to face their biggest workplace transmission due to AI agents. But nobody asks them what they want.

While AI races to automate everything, we took a different approach: auditing what workers want vs. what AI can do across the US workforce.🧵
Shirley Wu (@shirleyyxwu) 's Twitter Profile Photo

Even the smartest LLMs can fail at basic multiturn communication Ask for grocery help → without asking where you live 🤦‍♀️ Ask to write articles → assumes your preferences 🤷🏻‍♀️ ⭐️CollabLLM (top 1%; oral ICML Conference) transforms LLMs from passive responders into active collaborators.

Even the smartest LLMs can fail at basic multiturn communication

Ask for grocery help → without asking where you live 🤦‍♀️
Ask to write articles → assumes your preferences 🤷🏻‍♀️

⭐️CollabLLM (top 1%; oral <a href="/icmlconf/">ICML Conference</a>) transforms LLMs from passive responders into active collaborators.
Yutong Zhang (@zhangyt0704) 's Twitter Profile Photo

AI companions aren’t science fiction anymore 🤖💬❤️ Thousands are turning to AI chatbots for emotional connection – finding comfort, sharing secrets, and even falling in love. But as AI companionship grows, the line between real and artificial relationships blurs. 📰 “Can A.I.

AI companions aren’t science fiction anymore 🤖💬❤️
Thousands are turning to AI chatbots for emotional connection – finding comfort, sharing secrets, and even falling in love. But as AI companionship grows, the line between real and artificial relationships blurs.

📰 “Can A.I.
CLS (@chengleisi) 's Twitter Profile Photo

Are AI scientists already better than human researchers? We recruited 43 PhD students to spend 3 months executing research ideas proposed by an LLM agent vs human experts. Main finding: LLM ideas result in worse projects than human ideas.

Are AI scientists already better than human researchers?

We recruited 43 PhD students to spend 3 months executing research ideas proposed by an LLM agent vs human experts.

Main finding: LLM ideas result in worse projects than human ideas.