Sean Trott (@sean_trott) 's Twitter Profile
Sean Trott

@sean_trott

Assistant Professor in Cognitive Science and Computational Social Science at UC San Diego. He/him.

ID: 624915950

linkhttps://seantrott.github.io/ calendar_today02-07-2012 19:04:24

179 Tweet

312 Followers

82 Following

Tom McCoy (@rtommccoy) 's Twitter Profile Photo

🤖🧠New paper!🧠🤖 A common assumption: In neural networks, "inductive bias" = "model architecture" But in fact initial weights also greatly influence inductive bias! This makes it possible for the same bias to be realized in very different architectures - albeit w/ limits

Cameron Jones (@camrobjones) 's Twitter Profile Photo

New preprint: we evaluated LLMs in a 3-party Turing test (participants speak to a human & AI simultaneously and decide which is which). GPT-4.5 (when prompted to adopt a humanlike persona) was judged to be the human 73% of the time, suggesting it passes the Turing test (🧵)

New preprint: we evaluated LLMs in a 3-party Turing test (participants speak to a human & AI simultaneously and decide which is which).

GPT-4.5 (when prompted to adopt a humanlike persona) was judged to be the human 73% of the time, suggesting it passes the Turing test (🧵)
Sheridan Feucht (@sheridan_feucht) 's Twitter Profile Photo

[📄] Are LLMs mindless token-shifters, or do they build meaningful representations of language? We study how LLMs copy text in-context, and physically separate out two types of induction heads: token heads, which copy literal tokens, and concept heads, which copy word meanings.

[📄] Are LLMs mindless token-shifters, or do they build meaningful representations of language? We study how LLMs copy text in-context, and physically separate out two types of induction heads: token heads, which copy literal tokens, and concept heads, which copy word meanings.
Tal Linzen (@tallinzen) 's Twitter Profile Photo

This looks quite bad. I've interacted with a number of different NSF and NIH employees in the last couple of months. They're truly doing god's work right now. I'm extremely grateful for their work under these conditions. science.org/content/articl…

Philipp Schoenegger (@schoeneggerphil) 's Twitter Profile Photo

New preprint out with an amazing 40-person team! We find that Claude 3.5 Sonnet outperforms incentivised human persuaders in a >1000-participant live quiz-chat in deceptive and truthful directions!

New preprint out with an amazing 40-person team! We find that Claude 3.5 Sonnet outperforms incentivised human persuaders in a >1000-participant live quiz-chat in deceptive and truthful directions!
arlo_son (@gson_ai) 's Twitter Profile Photo

#NLProc AI Co-Scientists 🤖 can generate ideas, but can they spot mistakes? (not yet! 🚫) In my recent paper, we introduce SPOT, a dataset of STEM manuscripts (math, materials science, chemistry, physics, etc), annotated with real errors. SOTA models like o3, gemini-2.5-pro

#NLProc
AI Co-Scientists 🤖 can generate ideas, but can they spot mistakes? (not yet! 🚫)

In my recent paper, we introduce SPOT, a dataset of STEM manuscripts (math, materials science, chemistry, physics, etc), annotated with real errors.

SOTA models like o3, gemini-2.5-pro
Tom McCoy (@rtommccoy) 's Twitter Profile Photo

🤖🧠Paper out in Nature Communications! 🧠🤖 Bayesian models can learn rapidly. Neural networks can handle messy, naturalistic data. How can we combine these strengths? Our answer: Use meta-learning to distill Bayesian priors into a neural network! nature.com/articles/s4146… 1/n

🤖🧠Paper out in Nature Communications! 🧠🤖

Bayesian models can learn rapidly. Neural networks can handle messy, naturalistic data. How can we combine these strengths?

Our answer: Use meta-learning to distill Bayesian priors into a neural network!

nature.com/articles/s4146…

1/n
Michael Hu (@michahu8) 's Twitter Profile Photo

📢 today's scaling laws often don't work for predicting downstream task performance. For some pretraining setups, smooth and predictable scaling is the exception, not the rule. a quick read about scaling law fails: 📜arxiv.org/abs/2507.00885 🧵1/5👇

📢 today's scaling laws often don't work for predicting downstream task performance. For some pretraining setups, smooth and predictable scaling is the exception, not the rule.

a quick read about scaling law fails: 
📜arxiv.org/abs/2507.00885

🧵1/5👇
Cameron Jones (@camrobjones) 's Twitter Profile Photo

Incredibly excited to announce I’ll be starting as an Asst Professor in the Psychology Department at Stony Brook this fall! I’ll also be recruiting students this year so let me know if you know any students who might be interested!

Sigal Samuel (@sigalsamuel) 's Twitter Profile Photo

You may have heard examples of AI "scheming" against us — blackmailing, deceiving, etc. But do those represent the actual tendencies of the AI, or is the bad behavior showing up because researchers are strongly nudging that out of the AIs? Thoughts here: vox.com/future-perfect…

Sean Trott (@sean_trott) 's Twitter Profile Photo

Excited for #cogsci2025! - On 8/1, RivPam will be presenting a poster on our work on quantity representations in transformer models. - I’ll also be giving a talk that day on the role of language in human and machine intelligence.