
Andrew Saxe
@saxelab
Prof at @GatsbyUCL and @SWC_Neuro, trying to figure out how we learn.
Bluesky: @SaxeLab
Mastodon: @[email protected]
ID: 1193222240202035200
http://www.saxelab.org 09-11-2019 17:42:48
684 Tweet
5,5K Followers
380 Following

New paper with Leon and Erin Grant! Why do we see localized receptive fields so often, even in models without sparisity regularization? We present a theory in the minimal setting from Alessandro Ingrosso and Sebastian Goldt





Our paper, “Make Haste Slowly: A Theory of Emergent Structured Mixed Selectivity in Feature Learning ReLU Networks” will be presented at ICLR 2025 this week (openreview.net/forum?id=27SSn…)! We derive closed-form dynamics for some, remarkably linear, feature learning ReLU networks (1/9)


Our paper just came out in PRX! Congrats to Nishil Patel and the rest of the team. TL;DR : We analyse neural network learning through the lens of statistical physics, revealing distinct scaling regimes with sharp transitions. 🔗 journals.aps.org/prx/abstract/1…

Really happy to see this paper out, led by Nishil Patel in collaboration with Stefano Sarao Mannelli and Andrew Saxe: we apply the statistical physics toolbox to analyse a simple model of reinforcement learning, and find some cool effects, like a speed-accuracy trade-off for generalisation 🚀



How do task dynamics impact learning in networks with internal dynamics? Excited to share our ICML Oral paper on learning dynamics in linear RNNs! with Clémentine Dominé 🍊 Murray Shanahan Pedro Mediano openreview.net/forum?id=KGOcr…