Weinan Sun (@sunw37) 's Twitter Profile
Weinan Sun

@sunw37

Neuroscience, Artificial Intelligence, and Beyond.
Assistant professor, Neurobiology and Behavior @CornellNBB

ID: 702739245195096065

calendar_today25-02-2016 06:17:47

874 Tweet

779 Followers

612 Following

Aaditya Singh (@aaditya6284) 's Twitter Profile Photo

Transformers employ different strategies through training to minimize loss, but how do these tradeoff and why? Excited to share our newest work, where we show remarkably rich competitive and cooperative interactions (termed "coopetition") as a transformer learns. Read on 🔎⏬

Transformers employ different strategies through training to minimize loss, but how do these tradeoff and why?

Excited to share our newest work, where we show remarkably rich competitive and cooperative interactions (termed "coopetition") as a transformer learns.

Read on 🔎⏬
Kim Stachenfeld (neurokim.bsky.social) (@neuro_kim) 's Twitter Profile Photo

Want to procedurally generate large-scale relational reasoning experiments in natural language, to study human psychology 🧠 or eval LLMs 🤖? We have a tool for that! github.com/google-deepmin… Check out Kenneth Marino's thread for some stuff you can do:

Dan Levenstein (@dlevenstein) 's Twitter Profile Photo

At #Cosyne2025? Come by my poster today (47) to hear how sequential predictive learning produces a continuous neural manifold that can generate replay during sleep, and spatial representations that "sweep" ahead to future positions. All from sensory information + action!

Astera Institute (@asterainstitute) 's Twitter Profile Photo

.Chongxi Lai works at the intersection of neuroscience, AI, and brain-machine interfaces. He researches building brain-like models in a simulated environment, to test whether cognition can be enhanced through novel AI-assisted BMI closed-loop stimulation algorithms.

Chongxi Lai (@chongxilai) 's Twitter Profile Photo

Thrilled to announce I've joined Astera Institute's first residency cohort! Excited to collaborate with this amazing team to build technology for a brighter future! I will focus on building and testing brain-like model in large-scale simulation and use AI to enhance it!

Alex Kwan 關進晞 (@kwanalexc) 's Twitter Profile Photo

Our latest study identifies a specific cell type and receptor essential for psilocybin’s long-lasting neural and behavioral effects 🍄🔬🧠🐁 Led by Ling-Xiao Shao and Clara Liao Funded by NIH National Institute of Mental Health (NIMH) 📄Read in nature - nature.com/articles/s4158… 1/12

Thomas Miconi (@thomasmiconi) 's Twitter Profile Photo

New preprint! Intelligent creatures can solve truly novel problems (not just variations of previous problems), zero-shot! Why? They can "think" before acting, i.e. mentally simulate possible behaviors and evaluate likely outcomes How can we build agents with this ability?

Chen Sun 🤖🧠🇨🇦 (@chensun92) 's Twitter Profile Photo

From our team at Google DeepMind: we ask, as an LLM continues to learn, how do new facts pollute existing knowledge? (and can we control it) We investigate such hallucinations in our new paper, to be presented as Spotlight at ICLR 2026 next week.

From our team at <a href="/GoogleDeepMind/">Google DeepMind</a>: we ask, as an LLM continues to learn, how do new facts pollute existing knowledge?  (and can we control it)

We investigate such hallucinations in our new paper, to be presented as Spotlight at <a href="/iclr_conf/">ICLR 2026</a> next week.
Pierre Apostolides (@pfapostolides) 's Twitter Profile Photo

(Plz repost) I’ve been receiving some good news lately and will be hiring at all levels to expand the lab. Please get in contact if you are interested in reinforcement learning, neural plasticity, circuit dynamics, and/or hearing rehabilitation. pierre.apostolides @ umich .edu

Dileep George (@dileeplearning) 's Twitter Profile Photo

Cool work from HHMI | Janelia .. "cognitive graphs of latent structure" .... Looks like even more evidence for CSCG-like representation and schemas. (science.org/doi/10.1126/sc…, arxiv.org/abs/2302.07350) biorxiv.org/content/10.110…

Andrew Saxe (@saxelab) 's Twitter Profile Photo

How does in-context learning emerge in attention models during gradient descent training? Sharing our new Spotlight paper ICML Conference: Training Dynamics of In-Context Learning in Linear Attention arxiv.org/abs/2501.16265 Led by Yedi Zhang with Aaditya Singh and Peter Latham

Kevin Ellis (@ellisk_kellis) 's Twitter Profile Photo

New paper: World models + Program synthesis by Wasu Top Piriyakulkij 1. World modeling on-the-fly by synthesizing programs w/ 4000+ lines of code 2. Learns new environments from minutes of experience 3. Positive score on Montezuma's Revenge 4. Compositional generalization to new environments