Justus Mattern (@matternjustus) 's Twitter Profile
Justus Mattern

@matternjustus

Research Engineer @PrimeIntellect | prev. co-founder re.video (YC S23), research @MPI_IS, physics @RWTH

ID: 1371519617550663687

linkhttps://www.justusmattern.com calendar_today15-03-2021 17:52:22

570 Tweet

2,2K Followers

300 Following

Justus Mattern (@matternjustus) 's Twitter Profile Photo

great thread, summarizes well why we're particularly excited about RL higher inference to training compute ratio -> less inter node communication -> better suited for globally distributed training infra with slow connection speeds

Justus Mattern (@matternjustus) 's Twitter Profile Photo

SYNTHETIC-2 Datasets are now on Huggingface! We’re releasing an SFT dataset collected from the new R1-0528 as well as an RL Dataset with difficulty annotations from various smaller models. Go train some models 🫡

Jackmin (@jackminong) 's Twitter Profile Photo

Toploc Poster session tomorrow (Wed) at 4:30 PM East Hall E-1106 I’ll be around through Saturday; if you’re into decentralized training & inference, lets chat!

Toploc Poster session tomorrow (Wed) at 4:30 PM East Hall E-1106

I’ll be around through Saturday;  if you’re into decentralized training & inference, lets chat!
Simon Guo 🦝 (@simonguozirui) 's Twitter Profile Photo

At #ICML2025 in Vancouver 🇨🇦 this week, presenting some work from my first year at Stanford! Come find me at posters or just around the conference! Thursday: KernelBench: Can LLMs Write Efficient GPU Kernels? 11AM East E-2010 Saturday: Kevin: Multi-Turn RL for Generating

Mario Sieg (@_mario_neo_) 's Twitter Profile Photo

my piquant quantization kernels are almost 50 times faster than pytorch's on the CPU. pytorch’s sub‑byte quantization (torch.quint4x2, torch.quint2x4) is quite slow. 1/2

my piquant quantization kernels are almost 50 times faster than pytorch's on the CPU.

pytorch’s sub‑byte quantization (torch.quint4x2, torch.quint2x4) is quite slow.

1/2
Justus Mattern (@matternjustus) 's Twitter Profile Photo

RL with predefined tools does not matter in the long term, the most bitter lesson pilled approach is giving the model a single universal tool (a computer)

Minn (@minney_cat) 's Twitter Profile Photo

this year alone, I've met hundreds of the world's elite AI researchers + engineers steering the future of intelligence, and they ultimately want to do it here, in the US 🇺🇸 If a visa or green card are holding you back, join us on 7/31 in SF and hear real stories from

Justus Mattern (@matternjustus) 's Twitter Profile Photo

While LLMs are good at generating functionally correct frontend code, it’s stunning how bad AI-generated UIs are; I’m certain that this can become better with appropriate evals and reward signals Really excited about this leaderboard and the very hard-working team behind it!

Justus Mattern (@matternjustus) 's Twitter Profile Photo

Just landed in SF, I'm now stranded here without a desk while the rest of my team is still in Europe. If anyone can host me at their office for (one of) the next few days, please let me know (DMs open 🥹👉👈)

Chujie Zheng (@chujiezheng) 's Twitter Profile Photo

Noticed some curiosity about the specific score comparison between GSPO and GRPO. From our perspective, we’re more focused on scalability — can we achieve better performance by increasing compute (e.g., training with more steps, extending generation length, regularly updating