William Merrill (@lambdaviking) 's Twitter Profile
William Merrill

@lambdaviking

Will irl - Ph.D. student @NYUDataScience

ID: 391600060

linkhttps://lambdaviking.com/ calendar_today15-10-2011 20:39:17

1,1K Tweet

2,2K Followers

633 Following

Michael Hu (@michahu8) 's Twitter Profile Photo

Training on a little ๐Ÿค formal language BEFORE natural language can make pretraining more efficient! How and why does this work? The answer liesโ€ฆBetween Circuits and Chomsky. ๐Ÿงต1/6๐Ÿ‘‡

Training on a little ๐Ÿค formal language BEFORE natural language can make pretraining more efficient!

How and why does this work? The answer liesโ€ฆBetween Circuits and Chomsky.

๐Ÿงต1/6๐Ÿ‘‡
Kimon Fountoulakis (@kfountou) 's Twitter Profile Photo

Computational Capability and Efficiency of Neural Networks: A Repository of Papers I compiled a list of theoretical papers related to the computational capabilities of Transformers, recurrent networks, feedforward networks, and graph neural networks. Link:

Computational Capability and Efficiency of Neural Networks: A Repository of Papers

I compiled a list of theoretical papers related to the computational capabilities of Transformers, recurrent networks, feedforward networks, and graph neural networks.

Link:
Tal Linzen (@tallinzen) 's Twitter Profile Photo

International students, and Chinese students in particular, are essential to the AI research ecosystem in the US. You can't say you support AI research in this country and then threaten to revoke Chinese students' visas.

Byung-Doh Oh (@byungdoh) 's Twitter Profile Photo

Have reading time corpora been leaked into LM pre-training corpora? Should you be cautious about using pre-trained LM surprisal as a consequence? We identify the longest overlapping token sequences and conclude the leakage is mostly not severe. In Findings of #ACL2025 #ACL2025NLP

Have reading time corpora been leaked into LM pre-training corpora? Should you be cautious about using pre-trained LM surprisal as a consequence? We identify the longest overlapping token sequences and conclude the leakage is mostly not severe. In Findings of #ACL2025 #ACL2025NLP
Tal Linzen (@tallinzen) 's Twitter Profile Photo

Slides from my talk at Apple (thanks for hosting!) on our recent work on formal languages for LLM pretraining and evaluation: drive.google.com/file/d/1EtsyQ-โ€ฆ

Slides from my talk at Apple (thanks for hosting!) on our recent work on formal languages for LLM pretraining and evaluation: drive.google.com/file/d/1EtsyQ-โ€ฆ
Jackson Petty (@jowenpetty) 's Twitter Profile Photo

How well can LLMs understand tasks with complex sets of instructions? We investigate through the lens of RELIC: REcognizing (formal) Languages In-Context, finding a significant overhang between what LLMs are able to do theoretically and how well they put this into practice.

How well can LLMs understand tasks with complex sets of instructions? We investigate through the lens of RELIC: REcognizing (formal) Languages In-Context, finding a significant overhang between what LLMs are able to do theoretically and how well they put this into practice.
William Merrill (@lambdaviking) 's Twitter Profile Photo

A fun project with really thorough analysis of how LLMs try and often fail to implement parsing algorithms. Bonus: find out what this all has to do with the Kalamang language from New Guinea

William Merrill (@lambdaviking) 's Twitter Profile Photo

I'll be defending my dissertation at NYU next Monday, June 16 at 4pm ET! I've definitely missed inviting some people who might be interested, so please email me if you'd like to attend (NYC or Zoom)

I'll be defending my dissertation at NYU next Monday, June 16 at 4pm ET!

I've definitely missed inviting some people who might be interested, so please email me if you'd like to attend (NYC or Zoom)
Tal Linzen (@tallinzen) 's Twitter Profile Photo

So, on the topic of the Apple puzzle reasoning paper: we got pretty similar results in our recent paper on recognizing context-free languages as an LLM eval, a task that also requires the model to follow an algorithm (which I think is what LLM folks mean by "reasoning").

David Chiang (@davidweichiang) 's Twitter Profile Photo

New on arXiv: Knee-Deep in C-RASP, by Andy J Yang, Michael Cadilhac and me. The solid stepped line is our theoretical prediction based on what problems C-RASP can solve, and the numbers/colors are what transformers (no position embedding) can learn.

New on arXiv: Knee-Deep in C-RASP, by <a href="/pentagonalize/">Andy J Yang</a>, Michael Cadilhac and me. The solid stepped line is our theoretical prediction based on what problems C-RASP can solve, and the numbers/colors are what transformers (no position embedding) can learn.
TTIC (@ttic_connect) 's Twitter Profile Photo

Weโ€™re proud to announce three new tenure-track assistant professors joining TTIC in Fall 2026: Yossi Gandelsman (Yossi Gandelsman), Will Merrill (William Merrill), and Nick Tomlin (Nicholas Tomlin). Meet them here: buff.ly/JH1DFtT

Weโ€™re proud to announce three new tenure-track assistant professors joining TTIC in Fall 2026: Yossi Gandelsman (<a href="/YGandelsman/">Yossi Gandelsman</a>), Will Merrill (<a href="/lambdaviking/">William Merrill</a>), and Nick Tomlin (<a href="/NickATomlin/">Nicholas Tomlin</a>). Meet them here: buff.ly/JH1DFtT