Sean McLeish (@seanmcleish) 's Twitter Profile
Sean McLeish

@seanmcleish

PhD student at the University of Maryland

ID: 1727125615600451584

calendar_today22-11-2023 00:43:49

72 Tweet

432 Followers

97 Following

𝚐𝔪𝟾𝚡𝚡𝟾 (@gm8xx8) 's Twitter Profile Photo

Gemstones: A Model Suite for Multi-Faceted Scaling Laws paper: arxiv.org/abs/2502.06857 code: github.com/mcleish7/gemst… The project analyzes scaling laws across 22 AI models, ranging from 50M to 2B parameters, by varying model width and depth.

Tom Goldstein (@tomgoldsteincs) 's Twitter Profile Photo

Our new models for studying scaling laws are out! The Gemstones are 4K checkpoints (22 models) trained on 10T token combined, with varying architectures and learning rates. Here’s my fav new scaling experiment. It explains why industry has abandoned big dense models 🧵 (1/4)

AGI.Eth (@ceobillionaire) 's Twitter Profile Photo

Gemstones: A Model Suite for Multi-Faceted Scaling Laws McLeish et al.: arxiv.org/abs/2502.06857 #ArtificialIntelligence #DeepLearning #MachineLearning

Gemstones: A Model Suite for Multi-Faceted Scaling Laws

McLeish et al.: arxiv.org/abs/2502.06857

#ArtificialIntelligence #DeepLearning #MachineLearning
DAIR.AI (@dair_ai) 's Twitter Profile Photo

Here are the top AI Papers of the Week (Feb 10-16): - Latent Reasoning - Large Memory Models - Brain-to-Text Decoding - Enhancing Reasoning to Adapt LLMs - Reinforcement Learning via Self-Play - Competitive Programming with Large Reasoning Models Read on for more:

Sean McLeish (@seanmcleish) 's Twitter Profile Photo

If there is another McLeish/Schwarzschild duo doing recurrent deep learning out there, Avi Schwarzschild and I would love to meet you. Otherwise how are we dealing with the overload of ChatGPT made up references these days?! 🤯

If there is another McLeish/Schwarzschild duo doing recurrent deep learning out there, <a href="/A_v_i__S/">Avi Schwarzschild</a> and I would love to meet you. Otherwise how are we dealing with the overload of ChatGPT made up references these days?! 🤯
Ashwinee Panda (@pandaashwinee) 's Twitter Profile Photo

people are talking about whether scaling laws are broken or pretraining is saturating. so what does that even mean? consider the loss curves from our recent gemstones paper. as we add larger models, the convex hull doesn’t flatten out on this log-log plot. that's good!

people are talking about whether scaling laws are broken or pretraining is saturating. so what does that even mean? consider the loss curves from our recent gemstones paper. as we add larger models, the convex hull doesn’t flatten out on this log-log plot. that's good!
Alexander Doria (@dorialexander) 's Twitter Profile Photo

Multiple interesting experiments and findings for pretraining recipes. I especially liked the part about width/depth (not trivial to choose when you’re in the 1-3b range).

The TWIML AI Podcast (@twimlai) 's Twitter Profile Photo

Today, we're joined by Jonas Geiping, research group leader at ELLIS Institute Tübingen and the Intelligent Systems to discuss his recent paper, “Scaling up Test-Time Compute with Latent Reasoning: A Recurrent Depth Approach.” This paper proposes a novel language model architecture which uses

Lawrence Livermore National Laboratory (@livermore_lab) 's Twitter Profile Photo

#LLNL is not just advancing #AI, we are redefining how AI and science converge to unlock the next era of discovery. Huginn represents a new breed of language models that emphasizes careful introspection over immediate yet often incomplete answers. livermorelab.info/4ianNIM

Dayal Kalra (@dayal_kalra) 's Twitter Profile Photo

Excited to share our paper "Universal Sharpness Dynamics..." is accepted to #ICLR2025! Neural net training exhibits rich curvature (sharpness) dynamics (sharpness reduction, progressive sharpening, Edge of Stability)- but why?🤔 We show that a minimal model captures it all! 1/n

John Kirchenbauer (@jwkirchenbauer) 's Twitter Profile Photo

Before you leave Singapore be sure to check out the Tomlab's trio of pretraining papers at the Open Science for Foundation Models (SCI-FM) workshop in Hall 4 #5 ! Jonas and I will be around the rest of the afternoon to share our amd war stories 🥲

Before you leave Singapore be sure to check out the Tomlab's trio of pretraining papers at the Open Science for Foundation Models (SCI-FM) workshop in Hall 4 #5 ! Jonas and I will be around the rest of the afternoon to share our amd war stories 🥲
Kimon Fountoulakis (@kfountou) 's Twitter Profile Photo

Update: 14 empirical papers added. 1. Learning to Execute. Wojciech Zaremba, Ilya Sutskever 2. Neural Programmer-Interpreters.Scott Reed, Nando de Freitas 3. Neural Programmer: Inducing Latent Programs with Gradient Descent. Arvind Neelakantan, Quoc V. Le, Ilya Sutskever 4.

Alexander Panfilov (@kotekjedi_ml) 's Twitter Profile Photo

Stronger models need stronger attackers! 🤖⚔️ In our new paper we explore how attacker-target capability dynamics affect red-teaming success (ASR). Key insights: 🔸Stronger models = better attackers 🔸ASR depends on capability gap 🔸Psychology >> STEM for ASR More in 🧵👇

Stronger models need stronger attackers! 🤖⚔️
In our new paper we explore how attacker-target capability dynamics affect red-teaming success (ASR).

Key insights:
🔸Stronger models = better attackers
🔸ASR depends on capability gap
🔸Psychology &gt;&gt; STEM for ASR

More in 🧵👇
Ruchit Rawal (@rawalruchit) 's Twitter Profile Photo

Introducing ARGUS 👁️ A benchmark for measuring hallucinations and omissions in free-form captions generated by Video-LLMs.

Introducing ARGUS 👁️

A benchmark for measuring hallucinations and omissions in free-form captions generated by Video-LLMs.
Avi Schwarzschild (@a_v_i__s) 's Twitter Profile Photo

Ever tried to tell if someone really forgot your birthday? ... evaluating forgetting is tricky. Now imagine doing that… but for an LLM… with privacy on the line. We studied how to evaluate machine unlearning, and we found some problems. 🧵

Avi Schwarzschild (@a_v_i__s) 's Twitter Profile Photo

Big news! 🎉 I’m joining UNC-Chapel Hill as an Assistant Professor in Computer Science starting next year! Before that, I’ll be spending time OpenAI working on LLM privacy. UNC Computer Science UNC NLP

Big news! 🎉  I’m joining UNC-Chapel Hill as an Assistant Professor in Computer Science starting next year! Before that, I’ll be spending time <a href="/OpenAI/">OpenAI</a> working on LLM privacy.
<a href="/unccs/">UNC Computer Science</a> <a href="/uncnlp/">UNC NLP</a>