Andrew Saxe (@saxelab) 's Twitter Profile
Andrew Saxe

@saxelab

Prof at @GatsbyUCL and @SWC_Neuro, trying to figure out how we learn.
Bluesky: @SaxeLab
Mastodon: @[email protected]

ID: 1193222240202035200

linkhttp://www.saxelab.org calendar_today09-11-2019 17:42:48

684 Tweet

5,5K Followers

380 Following

Andrew Saxe (@saxelab) 's Twitter Profile Photo

New paper with Leon and Erin Grant! Why do we see localized receptive fields so often, even in models without sparisity regularization? We present a theory in the minimal setting from Alessandro Ingrosso and Sebastian Goldt

Blavatnik Awards (@blavatnikawards) 's Twitter Profile Photo

2025 @Blavatnikawards UK 🇬🇧 Finalist Andrew Saxe from UCL was featured on the BBC Science Focus Instant Genius Podcast on how AI is helping us understand how our brains🧠 learn: open.spotify.com/episode/6pcv55… and podcasts.apple.com/gb/podcast/how… Andrew Saxe @UCL

Aaditya Singh (@aaditya6284) 's Twitter Profile Photo

Transformers employ different strategies through training to minimize loss, but how do these tradeoff and why? Excited to share our newest work, where we show remarkably rich competitive and cooperative interactions (termed "coopetition") as a transformer learns. Read on 🔎⏬

Transformers employ different strategies through training to minimize loss, but how do these tradeoff and why?

Excited to share our newest work, where we show remarkably rich competitive and cooperative interactions (termed "coopetition") as a transformer learns.

Read on 🔎⏬
Clémentine Dominé 🍊 (@clementinedomi6) 's Twitter Profile Photo

Our paper, “A Theory of Initialization’s Impact on Specialization,” has been accepted to ICLR 2025! openreview.net/forum?id=RQz7s… We shows how neural network can build specialized and shared representation depending on initialization, this has consequences in continual learning. (1/8)

Our paper, “A Theory of Initialization’s Impact on Specialization,” has been accepted to ICLR 2025!
openreview.net/forum?id=RQz7s…
We shows how neural network can build specialized and shared representation depending on initialization, this has consequences in continual learning. 
(1/8)
Lenka Zdeborova (@zdeborova) 's Twitter Profile Photo

Happy to share the recording of my plenary talk at Cosyne 2025 two days ago. You will learn about the statistical physics approach, phase transitions in learning, transformers, sequence models, attention etc. youtube.com/watch?v=PurZcs…

Mohamady El-Gaby (@gabymohamady) 's Twitter Profile Photo

How do cognitive maps fail? And how can this help us understand/treat psychosis? My lab at Experimental Psychology, Oxford is hiring a Postdoc tinyurl.com/2p935hhz and RA tinyurl.com/3myfpb78 to answer these questions in mouse models. Here's why you might want to join: đź§µ

How do cognitive maps fail? And how can this help us understand/treat psychosis? My lab at Experimental Psychology, Oxford is hiring a Postdoc tinyurl.com/2p935hhz and RA tinyurl.com/3myfpb78 to answer these questions in mouse models.
Here's why you might want to join: đź§µ
Devon Jarvis (@devonjarvi5) 's Twitter Profile Photo

Our paper, “Make Haste Slowly: A Theory of Emergent Structured Mixed Selectivity in Feature Learning ReLU Networks” will be presented at ICLR 2025 this week (openreview.net/forum?id=27SSn…)! We derive closed-form dynamics for some, remarkably linear, feature learning ReLU networks (1/9)

Sebastian Goldt (@sebastiangoldt) 's Twitter Profile Photo

If I had known about this master when I was coming out of my Bachelor, I would have applied in a heartbeat, so please help me spread the word among the budding physicists with an interest in complex systems !

Stefano Sarao Mannelli (@stefsmlab) 's Twitter Profile Photo

Our paper just came out in PRX! Congrats to Nishil Patel and the rest of the team. TL;DR : We analyse neural network learning through the lens of statistical physics, revealing distinct scaling regimes with sharp transitions. 🔗 journals.aps.org/prx/abstract/1…

Sebastian Goldt (@sebastiangoldt) 's Twitter Profile Photo

Really happy to see this paper out, led by Nishil Patel in collaboration with Stefano Sarao Mannelli and Andrew Saxe: we apply the statistical physics toolbox to analyse a simple model of reinforcement learning, and find some cool effects, like a speed-accuracy trade-off for generalisation 🚀

Aaditya Singh (@aaditya6284) 's Twitter Profile Photo

Was super fun to be a part of this work! Felt very satisfying to bring the theory work on ICL with linear attention a bit closer to practice (with multi-headed low rank attention), and of course, add a focus on dynamics. Thread đź§µ with some extra highlights

Aaditya Singh (@aaditya6284) 's Twitter Profile Photo

Excited to share this work has been accepted as an Oral at #icml2025 -- looking forward to seeing everyone in Vancouver, and an extra thanks to my amazing collaborators for making this project so much fun to work on :)

Alexandra Proca (@a_proca) 's Twitter Profile Photo

How do task dynamics impact learning in networks with internal dynamics? Excited to share our ICML Oral paper on learning dynamics in linear RNNs! with Clémentine Dominé 🍊 Murray Shanahan Pedro Mediano openreview.net/forum?id=KGOcr…