Abdulkadir Gokce (@akgokce0) 's Twitter Profile
Abdulkadir Gokce

@akgokce0

IC SC @EPFL_en @ICepfl | EE&Math @unibogazici_en

ID: 1300695920405893122

calendar_today01-09-2020 07:26:44

12 Tweet

49 Followers

65 Following

Badr AlKhamissi (@bkhmsi) 's Twitter Profile Photo

🚨 New Paper!! How can we train LLMs using 100M words? In our babyLM paper, we introduce a new self-synthesis training recipe to tackle this question! 🍼💻 This was a fun project co-led by me, Yingtian Tang, Abdulkadir Gokce, w/ Hannes Mehrer & Martin Schrimpf 🧵⬇️

🚨 New Paper!!

How can we train LLMs using 100M words? In our <a href="/babyLMchallenge/">babyLM</a> paper, we introduce a new self-synthesis training recipe to tackle this question! 🍼💻

This was a fun project co-led by me, <a href="/yingtian80536/">Yingtian Tang</a>, <a href="/akgokce0/">Abdulkadir Gokce</a>, w/ <a href="/HannesMehrer/">Hannes Mehrer</a> &amp; <a href="/martin_schrimpf/">Martin Schrimpf</a>  

🧵⬇️
Badr AlKhamissi (@bkhmsi) 's Twitter Profile Photo

🚨 New Paper! Can neuroscience localizers uncover brain-like functional specializations in LLMs? 🧠🤖 Yes! We analyzed 18 LLMs and found units mirroring the brain's language, theory of mind, and multiple demand networks! w/ Greta Tuckute, Antoine Bosselut, & Martin Schrimpf 🧵👇

🚨 New Paper!

Can neuroscience localizers uncover brain-like functional specializations in LLMs? 🧠🤖

Yes! We analyzed 18 LLMs and found units mirroring the brain's language, theory of mind, and multiple demand networks!

w/ <a href="/GretaTuckute/">Greta Tuckute</a>, <a href="/ABosselut/">Antoine Bosselut</a>, &amp; <a href="/martin_schrimpf/">Martin Schrimpf</a> 

🧵👇
Badr AlKhamissi (@bkhmsi) 's Twitter Profile Photo

🚨 New Preprint!! LLMs trained on next-word prediction (NWP) show high alignment with brain recordings. But what drives this alignment—linguistic structure or world knowledge? And how does this alignment evolve during training? Our new paper explores these questions. 👇🧵

🚨 New Preprint!!

LLMs trained on next-word prediction (NWP) show high alignment with brain recordings. But what drives this alignment—linguistic structure or world knowledge? And how does this alignment evolve during training? Our new paper explores these questions. 👇🧵
Ben Lonnqvist (@lonnqvistben) 's Twitter Profile Photo

AI vision is insanely good nowadays—but is it really like human vision or something else entirely? In our new pre-print, we pinpoint a fundamental visual mechanism that's trivial for humans yet causes most models to fail spectacularly. Let's dive in👇🧠 [arxiv.org/abs/2504.05253]

AI vision is insanely good nowadays—but is it really like human vision or something else entirely? In our new pre-print, we pinpoint a fundamental visual mechanism that's trivial for humans yet causes most models to fail spectacularly. Let's dive in👇🧠
[arxiv.org/abs/2504.05253]
Badr AlKhamissi (@bkhmsi) 's Twitter Profile Photo

🚨New Preprint!! Thrilled to share with you our latest work: “Mixture of Cognitive Reasoners”, a modular transformer architecture inspired by the brain’s functional networks: language, logic, social reasoning, and world knowledge. 1/ 🧵👇

🚨New Preprint!!

Thrilled to share with you our latest work: “Mixture of Cognitive Reasoners”, a modular transformer architecture inspired by the brain’s functional networks: language, logic, social reasoning, and world knowledge.

1/ 🧵👇
Yingtian Tang (@yingtian80536) 's Twitter Profile Photo

🧠 NEW PREPRINT Many-Two-One: Diverse Representations Across Visual Pathways Emerge from A Single Objective biorxiv.org/content/10.110…

Hannes Mehrer (@hannesmehrer) 's Twitter Profile Photo

🧠 New preprint: we show that model-guided microstimulation can steer monkey visual behavior. Paper: arxiv.org/abs/2510.03684 🧵