Nathan Lambert (@natolambert) 's Twitter Profile
Nathan Lambert

@natolambert

Figuring out AI @allen_ai, open models, RLHF, fine-tuning, etc
DM me elsewhere (email). Limited replies.
Writes @interconnectsai
Writing rlhfbook.com

ID: 2939913921

linkhttps://www.interconnects.ai/ calendar_today24-12-2014 20:14:33

8,8K Tweet

48,48K Followers

819 Following

Brian Huang ✈️ ICLR (@brianryhuang) 's Twitter Profile Photo

it's surprising that this is the case, but frontier labs are actually *understaffed* there's always an immense amount of low-hanging fruit in the short-to-medium-term, and borderline not enough people (or compute) to do it, hence the always-on intense work schedule and ruthless

Nathan Lambert (@natolambert) 's Twitter Profile Photo

I'd put good money on this being an high-impact finetune of one of the large, Chinese MoE models. I'm very excited to see more companies able to train models that suit their needs. Bodes very well for the ecosystem that specific data is stronger than a bigger, general model.

Nathan Lambert (@natolambert) 's Twitter Profile Photo

I'm a total sucker for nice RL training scaling plots. They're very neglected vis-a-vis the much easier inference-time scaling plots.

I'm a total sucker for nice RL training scaling plots.
They're very neglected vis-a-vis the much easier inference-time scaling plots.
Nathan Lambert (@natolambert) 's Twitter Profile Photo

I feel strongly that, while I understand the challenges they're feeling to run this, that this is the wrong decision. What Arxiv is in practice versus what it is in reality is very different. In practice there are already moderation rules, but they're so minimally enforced (due