Berkeley AI Research (@berkeley_ai) 's Twitter Profile
Berkeley AI Research

@berkeley_ai

We're graduate students, postdocs, faculty and scientists at the cutting edge of artificial intelligence research.

ID: 891077171673931776

linkhttp://bair.berkeley.edu/ calendar_today28-07-2017 23:25:27

978 Tweet

202,202K Followers

304 Following

Sergey Levine (@svlevine) 's Twitter Profile Photo

Embodied chain of thought (ECoT) is a powerful tool to get VLAs to think through problems, but why does it work? In our new work, we analyze various lightweight ECoT-like strategies, including co-training, to see what is the "minimal" amount of reasoning that can boost VLAs 🧵👇

Serina Chang (@serinachang5) 's Twitter Profile Photo

Excited to have two papers accepted to ACL 2025 main! 🎉 1. ChatBench with jake hofman Ashton Anderson - we conduct a large-scale user study converting static benchmark questions into human-AI conversations, showing how benchmarks fail to predict human-AI outcomes.

Excited to have two papers accepted to ACL 2025 main! 🎉 

1. ChatBench with <a href="/jakehofman/">jake hofman</a> <a href="/ashton1anderson/">Ashton Anderson</a> - we conduct a large-scale user study converting static benchmark questions into human-AI conversations, showing how benchmarks fail to predict human-AI outcomes.
Dawn Song (@dawnsongtweets) 's Twitter Profile Photo

🔐 Frontier AI is reshaping cybersecurity, raising critical new questions: 🔍 What is its current impact? ⚖️ Who stands to benefit more—attackers or defenders? 🛡️ How can we mitigate the risks? Addressing these challenges requires coordinated efforts across AI & security

🔐 Frontier AI is reshaping cybersecurity, raising critical new questions:
🔍 What is its current impact?
⚖️ Who stands to benefit more—attackers or defenders?
🛡️ How can we mitigate the risks?

Addressing these challenges requires coordinated efforts across AI &amp; security
ICML Conference (@icmlconf) 's Twitter Profile Photo

Invited talked are announced. icml.cc/virtual/2025/e… Jon Kleinberg Pamela Samuelson Frauke Kreuter Anca Dragan Andreas Krause

Sandya Subramanian (@sandyaphd) 's Twitter Profile Photo

It was a pleasure to share my PhD work on measuring pain in patients under anesthesia using wearable devices with UC Joint Computational Precision Health Program! My lab continues this work by studying diseases using novel wearable devices and algorithms. See our website at subramanianlab.com!

Akshat Gupta (@akshatgupta57) 's Twitter Profile Photo

Just did a major revision to our paper on Lifelong Knowledge Editing!🔍 Key takeaway (+ our new title) - "Lifelong Knowledge Editing requires Better Regularization" Fixing this leads to consistent downstream performance! Tom Hartvigsen Ahmed Alaa Gopala Anumanchipalli Berkeley AI Research

Just did a major revision to our paper on Lifelong Knowledge Editing!🔍

Key takeaway (+ our new title) - "Lifelong Knowledge Editing requires Better Regularization"

Fixing this leads to consistent downstream performance!

<a href="/tom_hartvigsen/">Tom Hartvigsen</a> <a href="/_ahmedmalaa/">Ahmed Alaa</a> <a href="/GopalaSpeech/">Gopala Anumanchipalli</a> <a href="/berkeley_ai/">Berkeley AI Research</a>
Yun S. Song (@yun_s_song) 's Twitter Profile Photo

How can one efficiently simulate phylodynamics for populations with billions of individuals, as is typical in many applications, e.g., viral evolution and cancer genomics? In this work with Michael Celentano, W. DeWitt, & S. Prillo, we provide a solution. doi.org/10.1073/pnas.2… 1/n

How can one efficiently simulate phylodynamics for populations with billions of individuals, as is typical in many applications, e.g., viral evolution and cancer genomics? In this work with <a href="/mcelentano/">Michael Celentano</a>, W. DeWitt, &amp; S. Prillo, we provide a solution. doi.org/10.1073/pnas.2…
1/n
Sergey Levine (@svlevine) 's Twitter Profile Photo

Goal-conditioned RL (GCRL) is great - unsupervised, can use data (in offline mode), flexibility to define tasks at test time. But can we run GCRL on *language data*?? In our new work we show that language GCRL enables sophisticated test-time reasoning for interactive tasks! 🧵👇

Xuandong Zhao (@xuandongzhao) 's Twitter Profile Photo

🚀 Excited to share the most inspiring work I’ve been part of this year: "Learning to Reason without External Rewards" TL;DR: We show that LLMs can learn complex reasoning without access to ground-truth answers, simply by optimizing their own internal sense of confidence. 1/n

🚀 Excited to share the most inspiring work I’ve been part of this year:
 
"Learning to Reason without External Rewards"

TL;DR: We show that LLMs can learn complex reasoning without access to ground-truth answers, simply by optimizing their own internal sense of confidence. 1/n
Ademi Adeniji (@ademiadeniji) 's Twitter Profile Photo

Closed-loop robot policies directly from human interactions. No teleop, no robot data co-training, no RL, and no sim. Just Aria smart glasses. Everyday human data is passively scalable and a massively underutilized resource in robotics...More to come here in the coming weeks.

Kayo Yin (@kayo_yin) 's Twitter Profile Photo

Happy to announce the first workshop on Pragmatic Reasoning in Language Models — PragLM @ COLM 2025! 🧠🎉 How do LLMs engage in pragmatic reasoning, and what core pragmatic capacities remain beyond their reach? 🌐 sites.google.com/berkeley.edu/p… 📅 Submit by June 23rd

Younggyo Seo (@younggyoseo) 's Twitter Profile Photo

Excited to present FastTD3: a simple, fast, and capable off-policy RL algorithm for humanoid control -- with an open-source code to run your own humanoid RL experiments in no time! Thread below 🧵

Junhao (Bear) Xiong (@junhaobearxiong) 's Twitter Profile Photo

Guide your favorite protein generative model with experimental data? Meet ProteinGuide - a method to condition pre-trained models on properties without retraining. We validated it both in silico by guiding ProteinMPNN and ESM3 on 3 tasks and in vitro by engineering base editors.

Guide your favorite protein generative model with experimental data? Meet ProteinGuide - a method to condition pre-trained models on properties without retraining. We validated it both in silico by guiding ProteinMPNN and ESM3 on 3 tasks and in vitro by engineering base editors.
Pieter Abbeel (@pabbeel) 's Twitter Profile Photo

FastTD3: "Minimum innovation, maximum results" Not the paper we had planned to write, but one of the works I am most proud of. We wanted to make sure our baseline (TD3) was a very solid baseline, so we added a few things that are already known to help in RL (large,

Berkeley AI Research (@berkeley_ai) 's Twitter Profile Photo

Congratulations to BAIR students and faculty for their Best Paper Awards at the recently held #ICRA2025 in Atlanta. BAIR Researchers from Masayoshi Tomizuka's lab and the Berkeley DeepDrive Consortium won the Best Paper in Automation for their paper "Physics-Aware Robotic

Congratulations to BAIR students and faculty for their Best Paper Awards at the recently held #ICRA2025 in Atlanta.  

BAIR Researchers from Masayoshi Tomizuka's lab and the Berkeley DeepDrive Consortium won the Best Paper in Automation for their paper "Physics-Aware Robotic
Kevin Frans (@kvfrans) 's Twitter Profile Photo

Stare at policy improvement and diffusion guidance, and you may notice a suspicious similarity... We lay out an equivalence between the two, formalizing a simple technique (CFGRL) to improve performance across-the-board when training diffusion policies. arxiv.org/abs/2505.23458

Stare at policy improvement and diffusion guidance, and you may notice a suspicious similarity...

We lay out an equivalence between the two, formalizing a simple technique (CFGRL) to improve performance across-the-board when training diffusion policies.

arxiv.org/abs/2505.23458
Ritwik Gupta 🇺🇦 (@ritwik_g) 's Twitter Profile Photo

Ever wondered if the way we feed image patches to vision models is the best way? The standard row-by-row scan isn't always optimal! Modern long-sequence transformers can be surprisingly sensitive to patch order. We developed REOrder to find better, task-specific patch sequences.