Xavier Gonzalez (@xavierjgonzalez) 's Twitter Profile
Xavier Gonzalez

@xavierjgonzalez

PhD candidate in AI and Machine Learning at @Stanford. Advised by @scott_linderman. Parallelizing nonlinear RNNs. All views my own.

ID: 1353428973808566272

linkhttps://www.linkedin.com/in/xavier-gonzalez-517b5262/ calendar_today24-01-2021 19:49:52

75 Tweet

282 Followers

620 Following

Alex Wang (@heyyalexwang) 's Twitter Profile Photo

did you know you've been doing test-time learning this whole time? transformers, SSMs, RNNs, are all test-time regressors but with different design choices we present a unifying framework that derives sequence layers (and higher-order attention👀) from a *single* equation 🧵

did you know you've been doing test-time learning this whole time?

transformers, SSMs, RNNs, are all test-time regressors but with different design choices

we present a unifying framework that derives sequence layers (and higher-order attention👀) from a *single* equation

🧵
Xavier Gonzalez (@xavierjgonzalez) 's Twitter Profile Photo

Excited to share that our latest research on parallelizing nonlinear RNNS is being featured on the arXiv discussion forum alphaXiv. I will be on alphaXiv to answer any questions you have on the paper. alphaxiv.org/abs/2407.19115

Alex Wang (@heyyalexwang) 's Twitter Profile Photo

tomorrow at 10:30 pst/1:30 est i’ll be talking at the first ASAP seminar organized by Songlin Yang @ ICML '25 Simran Arora Xinyu Yang✈️ ICML 2025 Han Guo! i’ll present recent work on a unifying framework for current sequence models like mamba, attention, etc it’s all online so come thru!

Songlin Yang (@songlinyang4) 's Twitter Profile Photo

Recording: youtu.be/C7KnW8VFp4U Slides: asap-seminar.github.io/assets/slides/… If you're interested in this seminar series, please subscribe to our mailing list: groups.google.com/g/asap_seminar and join our Discord channel: discord.com/invite/vDaJTmK…

Xavier Gonzalez (@xavierjgonzalez) 's Twitter Profile Photo

This is an amazing class, a great way to learn cutting edge generative models like diffusion. And comes with a beautiful set of course notes diffusion.csail.mit.edu/docs/lecture-n… Tons of thanks to Peter Holderrieth for creating this super helpful resource!

Xavier Gonzalez (@xavierjgonzalez) 's Twitter Profile Photo

An incredible feature. The blog post it made for my paper "Towards Scalable and Stable Parallelization of nonlinear RNNs" was in some ways better than the blog post I put a lot of time into making! alphaxiv.org/overview/2407.… lindermanlab.github.io/hackathons/

Henry Zhong (@henryzhongsc) 's Twitter Profile Photo

Xavier Gonzalez NeurIPS Conference True this. Honestly, what is there to gain from having two deadlines a week apart? No reviewer action happens during that week, and we already know we'll be facing the heaviest submission load, where most reviewers are likely to max out. So having two separate PDFs (or worse, a

Xavier Gonzalez (@xavierjgonzalez) 's Twitter Profile Photo

I'm interning this summer at Apple doing machine learning research! I'll be in Seattle---would love to meet up with you if you are in town! Please reach out!

Kelly Buchanan (@ekellbuch) 's Twitter Profile Photo

LLMs can generate 100 answers, but which one is right? Check out our latest work closing the generation-verification gap by aggregating weak verifiers and distilling them into a compact 400M model. If this direction is exciting to you, we’d love to connect.

Rohan Paul (@rohanpaul_ai) 's Twitter Profile Photo

I asked ChatGPT (o3-pro) what the most unbelievable things it's learned about humans since being created was I find no-5 and the last one (meta-surprise) quite funny 🧵 Read on 👇 1. Simultaneous brilliance and self‑sabotage Humans can design spacecraft that navigate billions

I asked ChatGPT (o3-pro) what the most unbelievable things it's learned about humans since being created was

I find no-5 and the last one (meta-surprise) quite funny

🧵 Read on 👇

1. Simultaneous brilliance and self‑sabotage
Humans can design spacecraft that navigate billions
Jimmy Smith (@jimmysmith1919) 's Twitter Profile Photo

We are excited to release our first open-weight LFM models, optimized for on-device deployments. Extremely proud of the entire team! Check them out here: huggingface.co/LiquidAI

Keyon Vafa (@keyonv) 's Twitter Profile Photo

Can an AI model predict perfectly and still have a terrible world model? What would that even mean? Our new ICML paper formalizes these questions One result tells the story: A transformer trained on 10M solar systems nails planetary orbits. But it botches gravitational laws 🧵