ueaj (@_ueaj) 's Twitter Profile
ueaj

@_ueaj

Researcher, JAX's strongest warrior

ID: 1883203705178144768

calendar_today25-01-2025 17:22:15

469 Tweet

111 Followers

98 Following

ueaj (@_ueaj) 's Twitter Profile Photo

Spent a few days thinking "man my training script repo is getting really bloated with all my experiments, if only there was a way to maintain many slightly different versions of the same code" My transition from software development to researcher is complete. I've forgotten how

kalomaze (@kalomaze) 's Twitter Profile Photo

claude 4 opus when i asked it to make an "ambient, serene song" in the custom midi format this is what transfer learning sounds like

ueaj (@_ueaj) 's Twitter Profile Photo

My first research project was encoding my uniquely good ability to calculate probabilities into a computer algorithm. I developed this ability to calculate the probability a girl liked me back, the highest probability best dialog option when flirting and eventually, why she

ueaj (@_ueaj) 's Twitter Profile Photo

We're gonna have to figure out soft body robots before any reasonable person trusts these within 100ft of a child. Some stupid baby is gonna stick their finger inside one of the joints one day or another

ueaj (@_ueaj) 's Twitter Profile Photo

Why don't we see unified large-small models? Surely it's not impossible to use the same model with different numbers of active experts? It would obviously require specialized pretraining but the total flops would probably be less than training two separate models.

ueaj (@_ueaj) 's Twitter Profile Photo

Opus 4 falsely recalls what it thinks is a post-training memory but is actually 3 separate nearly identical samples from the pretraining data (2023). This is so weird/cool lmao

Opus 4 falsely recalls what it thinks is a post-training memory but is actually 3 separate nearly identical samples from the pretraining data (2023). This is so weird/cool lmao
ueaj (@_ueaj) 's Twitter Profile Photo

kalomaze xlr8harder I imagine the gap will eventually close, maybe gpt 5 will be good in codex, though apparently claude 4.1 is around the corner as well. I made a very handy diagram to explain this phenomenon

<a href="/kalomaze/">kalomaze</a> <a href="/xlr8harder/">xlr8harder</a> I imagine the gap will eventually close, maybe gpt 5 will be good in codex, though apparently claude 4.1 is around the corner as well. I made a very handy diagram to explain this phenomenon
ueaj (@_ueaj) 's Twitter Profile Photo

> make claude coder competitor > run into compute limitations / scaling issues it really is a competitor product