Ouail Kitouni (@wkitouni) 's Twitter Profile
Ouail Kitouni

@wkitouni

Member of technical staff @Anthropic prev @MIT @Meta @MSFTResearch

ID: 1159813920988762113

linkhttp://okitouni.github.io calendar_today09-08-2019 13:09:21

110 Tweet

65 Followers

95 Following

Ouail Kitouni (@wkitouni) 's Twitter Profile Photo

Repo to reproduce Grokking in a few lines of code (Full batch GD, small MLP, modular addition): github.com/okitouni/simpl…

Repo to reproduce Grokking in a few lines of code (Full batch GD, small MLP, modular addition): github.com/okitouni/simpl…
Summer Yue (@summeryue0) 's Twitter Profile Photo

🚀 Introducing the SEAL Leaderboards! We rank LLMs using private datasets that can’t be gamed. Vetted experts handle the ratings, and we share our methods in detail openly! Check out our leaderboards at scale.com/leaderboard! Which evals should we build next?

🚀 Introducing the SEAL Leaderboards! We rank LLMs using private datasets that can’t be gamed. Vetted experts handle the ratings, and we share our methods in detail openly! 

Check out our leaderboards at scale.com/leaderboard! 

Which evals should we build next?
Ouail Kitouni (@wkitouni) 's Twitter Profile Photo

You can just use a different model to prioritize higher signal tokens and generalize quicker. RHO-LOSS literally just works. (fraction is ratio of top-k tokens kept to total tokens; 1 is equiv to no rho-loss used)

You can just use a different model to prioritize higher signal tokens and generalize quicker. RHO-LOSS literally just works. (fraction is ratio of top-k tokens kept to total tokens; 1 is equiv to no rho-loss used)
Teortaxes▶️ (DeepSeek 推特🐋铁粉 2023 – ∞) (@teortaxestex) 's Twitter Profile Photo

Thesis from Ilya Sutskever : "to predict the next word, you have to predict the world" Antithesis from Yann LeCun : "AR-LLMs suck! Reversal curse!" Synthesis from FAIR: factorization-order-independent autoregressive model (MLM-U objective) (a paper subtweeting Mistral-PrefixLM thing?)

Thesis from <a href="/ilyasut/">Ilya Sutskever</a> : "to predict the next word, you have to predict the world"
Antithesis from <a href="/ylecun/">Yann LeCun</a> : "AR-LLMs suck! Reversal curse!"
Synthesis from FAIR: factorization-order-independent autoregressive model (MLM-U objective)

(a paper subtweeting Mistral-PrefixLM thing?)