
Andrea Boscutti
@aboscutti
ID: 1129888328915070976
18-05-2019 23:15:44
45 Tweet
31 Followers
299 Following


cognito Convolution is equivariant to translations. Self-attention is equivariant to permutations. They both have a role to play. Conv is efficient for signals with strong local correlations and motifs that can appear anywhere. SelfAtt is good for "object-based" representations where






Today we're excited to introduce Devin, the first AI software engineer. Devin is the new state-of-the-art on the SWE-Bench coding benchmark, has successfully passed practical engineering interviews from leading AI companies, and has even completed real jobs on Upwork. Devin is

Do LLMs really need to be so L? That's a rejected title for a new paper w/ Andrey Gromov, Kushal Tirumala, Hassan Shapourian, Paolo Glorioso on pruning open-weight LLMs: we can remove up to *half* the layers of Llama-2 70B w/ essentially no impact on performance on QA benchmarks. 1/




Best tiled matmul animation I've found on the internet. Thanks Michal Sojka (@[email protected])




Join us in congratulating the 2024 ACNP Travel Award Class! bit.ly/46GyJK1 Amanda Arulpragasam Estef Azevedo Igor D. Bandeira, MD, PhD Andrea Boscutti @CarinaSBrown Brenda Cabrera Mendoza Robert Y. Chen Seungwon (Sebastian) Choi 최승원 Dr. Kirstie Cummings Ashley Cunningham, MSc 🧠 Pasha Davoudian Kevin Dowling 🧬🧠 Lana Ruvolo, PhD



In the physical world, almost all information is transmitted through traveling waves -- why should it be any different in your neural network? Super excited to share recent work with the brilliant Mozes Jacobs: "Traveling Waves Integrate Spatial Information Through Time" 1/14
