Emiliano Penaloza (@emilianopp_) 's Twitter Profile
Emiliano Penaloza

@emilianopp_

PhD Student @Mila_Quebec

Interested in developing AI systems that prioritize user-focused and interpretable decision-making.

ID: 1841991137562595328

linkhttps://emilianopp.com calendar_today03-10-2024 23:58:19

31 Tweet

32 Followers

178 Following

Emmett Shear (@eshear) 's Twitter Profile Photo

Once again I am recommending that you NOT optimize your intelligent agents around usage metrics. Please. Pretty please. Just stop. I know that's how you optimize other products like recommendation engines but these things are NOT THE SAME.

Hafez Ghaemi (@hafezghm) 's Twitter Profile Photo

🚨 Preprint Alert 🚀 📄 seq-JEPA: Autoregressive Predictive Learning of Invariant-Equivariant World Models arxiv.org/abs/2505.03176 Can we simultaneously learn both transformation-invariant and transformation-equivariant representations with self-supervised learning (SSL)?

Siddarth Venkatraman (@siddarthv66) 's Twitter Profile Photo

Traj level objectives (RLOO, GRPO) seem better than value function based methods for LLM training (Q-learning, PPO). Quite unfortunate. Traj level RL can't do skill stitching, this is extremely sample inefficient for learning compositional reasoning (such as multi turn tool use).

Siddarth Venkatraman (@siddarthv66) 's Twitter Profile Photo

Is there a universal strategy to turn any generative model—GANs, VAEs, diffusion models, or flows—into a conditional sampler, or finetuned to optimize a reward function? Yes! Outsourced Diffusion Sampling (ODS) accepted to ICML Conference , does exactly that!

Is there a universal strategy to turn any generative model—GANs, VAEs, diffusion models, or flows—into a conditional sampler, or finetuned to optimize a reward function?
Yes! Outsourced Diffusion Sampling (ODS) accepted to <a href="/icmlconf/">ICML Conference</a> , does exactly that!
Benjamin Thérien (@benjamintherien) 's Twitter Profile Photo

Is AdamW the best inner optimizer for DiLoCo? Does the inner optimizer affect the compressibility of the DiLoCo delta? Excited to introduce MuLoCo: Muon is a practical inner optimizer for DiLoCo! 🧵arxiv.org/abs/2505.23725 1/N

Is AdamW the best inner optimizer for DiLoCo? Does the inner optimizer affect the compressibility of the DiLoCo delta? Excited to introduce MuLoCo: Muon is a practical inner optimizer for DiLoCo! 🧵arxiv.org/abs/2505.23725 1/N
Avery Ryoo (@averyryoo) 's Twitter Profile Photo

Super stoked to share my first first-author paper that introduces a hybrid architecture approach for real-time neural decoding. It's been a lot of work, but happy to showcase some very cool results!

Majdi Hassan (@majdi_has) 's Twitter Profile Photo

(1/n)🚨You can train a model solving DFT for any geometry almost without training data!🚨 Introducing Self-Refining Training for Amortized Density Functional Theory — a variational framework for learning a DFT solver that predicts the ground-state solutions for different

Luke Rowe (@luke22r) 's Twitter Profile Photo

🚀 Our method, Poutine, was the best-performing entry in the 2025 Waymo Vision-based End-to-End Driving Challenge at #CVPR2025! Our 3 B-parameter VLM Poutine scored 7.99 RFS on the official test set—comfortably ahead of every other entry (see figure).

🚀 Our method, Poutine, was the best-performing entry in the 2025 Waymo Vision-based End-to-End Driving Challenge at #CVPR2025!

Our 3 B-parameter VLM Poutine scored 7.99 RFS on the official test set—comfortably ahead of every other entry (see figure).