Alex Atanasov
@abatanasov
Fascinated by scaling and universality. PhD grad from @harvardphysics. Not a Bayesian. 🇧🇬 Opinions my own
ID: 823798863656054784
http://ABAtanasov.com 24-01-2017 07:45:29
327 Tweet
1,1K Followers
1,1K Following
Leo Gao jack morris This is demonstrably false. Neural nets in the kernel regime achieve generalization, overfit benignly, undergo double descent, and exhibit neural scaling laws (with e.g. chinchilla optimal compute predictions). Yes, feature learning is lacking, which means these miss transfer
as a linguist, the speaker should know better: it’s not about “speaking more efficiently” (using mathematical terms for prosaic things isn’t more efficient!) but rather to signal ingroup membership in an emerging elite technocrat class defined by it’s relation to math and code