Byung-Doh Oh (@byungdoh) 's Twitter Profile
Byung-Doh Oh

@byungdoh

ID: 49181595

linkhttps://byungdoh.github.io calendar_today21-06-2009 01:11:50

62 Tweet

88 Followers

97 Following

Michael Hu (@michahu8) 's Twitter Profile Photo

Training on a little 🤏 formal language BEFORE natural language can make pretraining more efficient! How and why does this work? The answer lies…Between Circuits and Chomsky. 🧵1/6👇

Training on a little 🤏 formal language BEFORE natural language can make pretraining more efficient!

How and why does this work? The answer lies…Between Circuits and Chomsky.

🧵1/6👇
NYU Center for Data Science (@nyudatascience) 's Twitter Profile Photo

CDS is hiring a Clinical Professor of Data Science. Teach ML, programming, and specialized courses in our 60 5th Ave building. Renewable contracts with promotion opportunities. Apply by April 1, 2025. For details, see: apply.interfolio.com/155349 #MachineLearning #ML #AIjobs

CDS is hiring a Clinical Professor of Data Science.

Teach ML, programming, and specialized courses in our 60 5th Ave building.

Renewable contracts with promotion opportunities.

Apply by April 1, 2025.

For details, see: apply.interfolio.com/155349

#MachineLearning #ML #AIjobs
Cory Shain (@coryshain) 's Twitter Profile Photo

New brain/language study w/ Ev (like in 'evidence', not Eve) Fedorenko 🇺🇦! We applied task-agnostic individualized functional connectomics (iFC) to the entire history of fMRI in the Fedorenko lab, parcellating nearly 1200 brains into networks based on activity fluctuations alone. doi.org/10.1101/2025.0… . 🧵

Jaap Jumelet (@jumeletj) 's Twitter Profile Photo

✨New paper ✨ Introducing 🌍MultiBLiMP 1.0: A Massively Multilingual Benchmark of Minimal Pairs for Subject-Verb Agreement, covering 101 languages! We present over 125,000 minimal pairs and evaluate 17 LLMs, finding that support is still lacking for many languages. 🧵⬇️

✨New paper ✨  
Introducing 🌍MultiBLiMP 1.0: A Massively Multilingual Benchmark of Minimal Pairs for Subject-Verb Agreement, covering 101 languages!  

We present over 125,000 minimal pairs and evaluate 17 LLMs, finding that support is still lacking for many languages. 

🧵⬇️
Ryo Yoshida (@yoshiryo0617) 's Twitter Profile Photo

Two papers have been accepted to ACL 2025 (one as first author). – Developmentally-plausible Working Memory Shapes… arxiv.org/abs/2502.04795 – If Attention Serves as a Cognitive Model of Human Memory Retrieval… arxiv.org/abs/2502.11469 #ACL2025 #ACL2025NLP

Two papers have been accepted to ACL 2025 (one as first author).
– Developmentally-plausible Working Memory Shapes… arxiv.org/abs/2502.04795
– If Attention Serves as a Cognitive Model of Human Memory Retrieval… arxiv.org/abs/2502.11469

#ACL2025 #ACL2025NLP
Tiago Pimentel (@tpimentelms) 's Twitter Profile Photo

A string may get 17 times less probability if tokenised as two symbols (e.g., ⟨he, llo⟩) than as one (e.g., ⟨hello⟩)—by an LM trained from scratch in each situation! Our #acl2025nlp paper proposes an observational method to estimate this causal effect! Longer thread soon!

A string may get 17 times less probability if tokenised as two symbols (e.g., ⟨he, llo⟩) than as one (e.g., ⟨hello⟩)—by an LM trained from scratch in each situation! Our #acl2025nlp paper proposes an observational method to estimate this causal effect! Longer thread soon!