Samuel Schmidgall (@srschmidgall) 's Twitter Profile
Samuel Schmidgall

@srschmidgall

PhD @JohnsHopkins // student researcher @Google @GoogleDeepmind // prev intern @AMD @Stanford

ID: 1323123590020169728

linkhttps://samuelschmidgall.github.io/ calendar_today02-11-2020 04:43:47

520 Tweet

2,2K Followers

413 Following

Google DeepMind (@googledeepmind) 's Twitter Profile Photo

We’re helping robots self-improve with the power of LLMs. 🤖 Introducing the Summarize, Analyze, Synthesize (SAS) prompt, which analyzes how they perform tasks based on previous actions and then suggests ways for them to get better using the medium of table tennis. 🏓

Khaled Saab (@khaledsaab11) 's Twitter Profile Photo

Gemini powers our multimodal health research! 💙 In our new paper on multimodal AMIE, we're pushing conversational diagnostic AI beyond text to handle images such as skin photos, ECGs, and clinical docs, which provide crucial context in healthcare. Blog: goo.gle/42D0QcB

Sundar Pichai (@sundarpichai) 's Twitter Profile Photo

What a finish! Gemini 2.5 Pro just completed Pokémon Blue!  Special thanks to Joel Z for creating and running the livestream, and to everyone who cheered Gem on along the way.

Google DeepMind (@googledeepmind) 's Twitter Profile Photo

Introducing AlphaEvolve: a Gemini-powered coding agent for algorithm discovery. It’s able to: 🔘 Design faster matrix multiplication algorithms 🔘 Find new solutions to open math problems 🔘 Make data centers, chip design and AI training more efficient across Google. 🧵

Max Fu (@letian_fu) 's Twitter Profile Photo

Tired of teleoperating your robots? We built a way to scale robot datasets without teleop, dynamic simulation, or even robot hardware. Just one smartphone scan + one human hand demo video → thousands of diverse robot trajectories. Trainable by diffusion policy and VLA models

Google DeepMind (@googledeepmind) 's Twitter Profile Photo

Deep Think in 2.5 Pro has landed. 🤯 It’s a new enhanced reasoning mode using our research in parallel thinking techniques - meaning it explores multiple hypotheses before responding. This enables it to handle incredibly complex math and coding problems more effectively.

Google DeepMind (@googledeepmind) 's Twitter Profile Photo

We’ve developed Gemini Diffusion: our state-of-the-art text diffusion model. Instead of predicting text directly, it learns to generate outputs by refining noise, step-by-step. This helps it excel at coding and math, where it can iterate over solutions quickly. #GoogleIO

Google Health (@googlehealth) 's Twitter Profile Photo

Exciting news from #GoogleIO! We’re introducing MedGemma, our most capable open model for multimodal medical text and image comprehension built on Gemma 3.

Shan Chen (@shan23chen) 's Twitter Profile Photo

Designing a hard but useful benchmark has always been a passion of mine. Here we present MedBrowseComp, a deep research + computer use benchmark that is easy to verify (like BrowseComp from OpenAI) but still very expandable 💊! Project page: moreirap12.github.io/mbc-browse-app/ 1/n

Designing a hard but useful benchmark has always been a passion of mine. Here we present MedBrowseComp, a deep research + computer use benchmark that is easy to verify (like BrowseComp from <a href="/OpenAI/">OpenAI</a>) but still very expandable 💊!

Project page:
moreirap12.github.io/mbc-browse-app/

1/n
Samuel Schmidgall (@srschmidgall) 's Twitter Profile Photo

🏥🤖 Autonomous surgical robots are expected to be capable of higher levels of autonomy due to the increased capabilities of AI Check out our new article in IEEE Spectrum which discusses the history and future of autonomous robotic surgery!

🏥🤖 Autonomous surgical robots are expected to be capable of higher levels of autonomy due to the increased capabilities of AI

Check out our new article in <a href="/IEEESpectrum/">IEEE Spectrum</a> which discusses the history and future of autonomous robotic surgery!
merve (@mervenoyann) 's Twitter Profile Photo

Google released MedGemma on I/O'25 👏 > 4B and 27B instruction fine-tuned vision LMs and a 4B pre-trained vision LM for medicine > available with transformers from the get-go 🤗 they also released a cool demo for scan reading ⤵️

Derya Unutmaz, MD (@deryatr_) 's Twitter Profile Photo

Benchmarks for AI agents in medicine are critically important! This new Deep Research benchmark looks very promising, similar to the recently announced HealthBench from OpenAI. Such benchmarks will significantly accelerate the development of AI doctors in the very near future!

Danielle Bitterman, MD (@dbittermanmd) 's Twitter Profile Photo

Agents are all the rage and we need to measure and track their abilities in the medical domain. Enter MedBrowseComp, the 1st benchmark to assess agents' abilities to reason, navigate the web, and search for verifiable med info! More👇 moreirap12.github.io/mbc-browse-app/

Alec Stapp (@alecstapp) 's Twitter Profile Photo

Reminder: Only 14% of US residents are immigrants. But immigrants are responsible for 36% of aggregate innovation. Two-thirds of this contribution is due to making their native-born collaborators better.

Reminder:

Only 14% of US residents are immigrants.

But immigrants are responsible for 36% of aggregate innovation.

Two-thirds of this contribution is due to making their native-born collaborators better.
DAIR.AI (@dair_ai) 's Twitter Profile Photo

8. MedBrowseComp MedBrowseComp is a new benchmark designed to evaluate LLM agents’ ability to perform complex, multi-hop medical fact-finding by browsing real-world, domain-specific web resources. x.com/shan23chen/sta…

Andrew Ng (@andrewyng) 's Twitter Profile Photo

I am alarmed by the proposed cuts to U.S. funding for basic research, and the impact this would have for U.S. competitiveness in AI and other areas. Funding research that is openly shared benefits the whole world, but the nation it benefits most is the one where the research is

Google DeepMind (@googledeepmind) 's Twitter Profile Photo

Introducing MedGemma, our most capable open model for multimodal medical text and image comprehension. 🩻 MedGemma is available now as part of Health AI Developer Foundations → goo.gle/medgemma

Michael Moor (@michael_d_moor) 's Twitter Profile Photo

🧵1/ ✨New preprint ✨ LLMs are getting better at answering medical questions. However, they still struggle to spot and fix errors in their own reasoning. That’s a big problem in medicine, where stakes are high and mistakes at any step could be critical. To address this issue,

🧵1/

✨New preprint ✨
LLMs are getting better at answering medical questions.
However, they still struggle to spot and fix errors in their own reasoning.

That’s a big problem in medicine, where stakes are high and mistakes at any step could be critical.

To address this issue,