Dara Bahri (@dara_bahri) 's Twitter Profile
Dara Bahri

@dara_bahri

Research Scientist at Google DeepMind.

ID: 1250475209959698432

linkhttp://www.dara.run calendar_today15-04-2020 17:25:03

83 Tweet

806 Followers

56 Following

Kevin Patrick Murphy (@sirbayes) 's Twitter Profile Photo

While GenAI is fun, I think its economic value is grossly over estimated, because it’s unreliable, risky and expensive to make and serve. It’s fine for creative tasks, but not (yet) autonomous agents

Christina Baek (@_christinabaek) 's Twitter Profile Photo

Did you know that the optimizer Sharpness Aware Minimization (SAM) is very robust to heavy label noise, with gains tens of percent above SGD? In our new work, we deep dive into how SAM achieves these gains. As it turns out, it’s not at all about sharpness at convergence!

Did you know that the optimizer Sharpness Aware Minimization (SAM) is very robust to heavy label noise, with gains tens of percent above SGD?

In our new work, we deep dive into how SAM achieves these gains. As it turns out, it’s not at all about sharpness at convergence!
Dara Bahri (@dara_bahri) 's Twitter Profile Photo

Driving endless stretches of hwy1 to run thru redwoods, waterfalls, up mtns, and swim in the pacific - complete freedom and the feeling that the future has infinite possibilities that is yours to take. Few things embody the American spirit, the American ethos better. Happy 4th!

Dara Bahri (@dara_bahri) 's Twitter Profile Photo

Back to Vienna for ICML, Mon evening to Fri evening. DM me if you're around and want to chat. And stop by our poster! arxiv.org/abs/2406.03682 (Tues poster session 2; also at high dim learning workshop on Fri)

Dara Bahri (@dara_bahri) 's Twitter Profile Photo

Want to watermark the outputs of an LLM but only have 3p API access to it? We propose a "distortion-free" LLM watermarking scheme that only requires a way to sample responses for the prompt. Joint work with John Wieting and others Google DeepMind (arxiv.org/abs/2410.02099).

John Wieting (@johnwieting2) 's Twitter Profile Photo

Check out our new paper1 We propose a watermarking scheme for LLMs that works with black-box access, is distortion-free, has provable guarentees, and can be chained with secret keys. Able to outperform white-box methods in comprehensive experiments. arxiv.org/abs/2410.02099.

Eric Jang (@ericjang11) 's Twitter Profile Photo

IIRC, This project began as mostly one researcher’s (Alex) interest at Google Brain. Nearly a decade later, it is a full fledged company that has made technology to transport scents digitally. A role model of patient research and dogged pursuit of a mission!

Fazl Barez (@fazlbarez) 's Twitter Profile Photo

🚨 New Paper Alert: Open Problem in Machine Unlearning for AI Safety 🚨 Can AI truly "forget"? While unlearning promises data removal, controlling emergent capabilities is a inherent challenge. Here's why it matters: 👇 Paper: arxiv.org/pdf/2501.04952 1/8

🚨 New Paper Alert: Open Problem in Machine Unlearning for AI Safety 🚨

Can AI truly "forget"? While unlearning promises data removal, controlling emergent capabilities is a inherent challenge. Here's why it matters: 👇

Paper: arxiv.org/pdf/2501.04952
1/8
Dara Bahri (@dara_bahri) 's Twitter Profile Photo

Really fun work with Richard Song! TLDR; we show that sequence modeling of numeric strings can be a powerful paradigm for tasks like regression and density estimation.

Dara Bahri (@dara_bahri) 's Twitter Profile Photo

"Mountains are not stadiums where I satisfy my ambition to achieve, they are the cathedrals where I practice my religion....I go to them as humans go to worship. From their lofty summits I view my past, dream of the future and, with an unusual acuity, am allowed to experience the

"Mountains are not stadiums where I satisfy my ambition to achieve, they are the cathedrals where I practice my religion....I go to them as humans go to worship. From their lofty summits I view my past, dream of the future and, with an unusual acuity, am allowed to experience the
Thomas Wolf (@thom_wolf) 's Twitter Profile Photo

After 6+ months in the making and burning over a year of GPU compute time, we're super excited to finally release the "Ultra-Scale Playbook" Check it out here: hf.co/spaces/nanotro… A free, open-source, book to learn everything about 5D parallelism, ZeRO, fast CUDA kernels,

After 6+ months in the making and burning over a year of GPU compute time, we're super excited to finally release the "Ultra-Scale Playbook"

Check it out here: hf.co/spaces/nanotro…

A free, open-source, book to learn everything about 5D parallelism, ZeRO, fast CUDA kernels,
Yasaman Bahri (@yasamanbb) 's Twitter Profile Photo

Excited to be at the APS March Meeting this year! American Physical Society I'll be giving a talk in the Tues afternoon session MAR-J58, Physics of Learning & Adaptation I.

Excited to be at the APS March Meeting this year! <a href="/APSphysics/">American Physical Society</a>

I'll be giving a talk in the Tues afternoon session MAR-J58, Physics of Learning &amp; Adaptation I.
Richard Song @ ICLR 2025 (@xingyousong) 's Twitter Profile Photo

Exactly why we spent last year on LLM regressors -> reward models to simulate expensive world feedback, not just humans. "To discover new ideas that go far beyond existing human knowledge, it is instead necessary to use grounded rewards: signals that arise from the environment

Exactly why we spent last year on LLM regressors -&gt; reward models to simulate expensive world feedback, not just humans.

"To discover new ideas that go far beyond existing human knowledge, it is instead necessary to use grounded rewards: signals that arise from the environment