Alisa Liu (@alisawuffles) 's Twitter Profile
Alisa Liu

@alisawuffles

PhD student at @uwcse @uwnlp

ID: 1197247629136203776

linkhttps://alisawuffles.github.io/ calendar_today20-11-2019 20:17:38

324 Tweet

2,2K Followers

350 Following

Taylor Sorensen (@ma_tay_) 's Twitter Profile Photo

🤔🤖Most AI systems assume there’s just one right answer—but many tasks have reasonable disagreement. How can we better model human variation? 🌍✨ We propose modeling at the individual-level using open-ended, textual value profiles! 🗣️📝 arxiv.org/abs/2503.15484 (1/?)

🤔🤖Most AI systems assume there’s just one right answer—but many tasks have reasonable disagreement. How can we better model human variation? 🌍✨

We propose modeling at the individual-level using open-ended, textual value profiles! 🗣️📝

arxiv.org/abs/2503.15484
(1/?)
Abhilasha Ravichander (@lasha_nlp) 's Twitter Profile Photo

Want to know what training data has been memorized by models like GPT-4? We propose information-guided probes, a method to uncover memorization evidence in *completely black-box* models, without requiring access to 🙅‍♀️ Model weights 🙅‍♀️ Training data 🙅‍♀️ Token probabilities 🧵1/5

Want to know what training data has been memorized by models like GPT-4?

We propose information-guided probes, a method to uncover memorization evidence in *completely black-box* models,

without requiring access to
🙅‍♀️ Model weights
🙅‍♀️ Training data
🙅‍♀️ Token probabilities 🧵1/5
Creston Brooks (@crestonbrooks) 's Twitter Profile Photo

Such a cool paper! Whitespace as a universal token delimiter = pretty arbitrary when there is little consensus what a "word" even is (esp. when you can save on inference)... there are counterexamples to any combination of criteria posed so far, e.g.: degruyter.com/document/doi/1…

Pratyush Maini (@pratyushmaini) 's Twitter Profile Photo

This is such a cool and intuitive modification to tokenization! And the results look just amazing both in terms of quality and inference speed.

Xenova (@xenovacom) 's Twitter Profile Photo

Love this! 🤗 SuperBPE is a *superword* tokenizer, which can encode multiple words using a single token (up to 33% more efficient than before)! 🤯 Plus, their official playground uses Transformers.js for in-browser tokenization and visualization! 🚀 Give it a try! 👇

Love this! 🤗 SuperBPE is a *superword* tokenizer, which can encode multiple words using a single token (up to 33% more efficient than before)! 🤯

Plus, their official playground uses Transformers.js for in-browser tokenization and visualization! 🚀

Give it a try! 👇
Etash Guha @ ICLR (@etash_guha) 's Twitter Profile Photo

Turns out, it’s possible to outperform DeepSeekR1-32B with only SFT on open data and no RL: Announcing OpenThinker2-32B and OpenThinker2-7B. We also release the data, OpenThoughts2-1M, curated by selecting quality instructions from diverse sources. 🧵 (1/n)

Turns out, it’s possible to outperform DeepSeekR1-32B with only SFT on open data and no RL: Announcing OpenThinker2-32B and OpenThinker2-7B. We also release the data, OpenThoughts2-1M, curated by selecting quality instructions from diverse sources. 🧵 (1/n)
Gonçalo Faria (@goncalorafaria) 's Twitter Profile Photo

Introducing 𝗤𝗔𝗹𝗶𝗴𝗻🚀, a 𝘁𝗲𝘀𝘁-𝘁𝗶𝗺𝗲 𝗮𝗹𝗶𝗴𝗻𝗺𝗲𝗻𝘁 𝗺𝗲𝘁𝗵𝗼𝗱 that improves language model performance using Markov chain Monte Carlo. With no model retraining, 𝗤𝗔𝗹𝗶𝗴𝗻 outperforms DPO-tuned models even when allowed to match inference compute, and achieves

Introducing 𝗤𝗔𝗹𝗶𝗴𝗻🚀, a 𝘁𝗲𝘀𝘁-𝘁𝗶𝗺𝗲 𝗮𝗹𝗶𝗴𝗻𝗺𝗲𝗻𝘁 𝗺𝗲𝘁𝗵𝗼𝗱 that improves language model performance using Markov chain Monte Carlo. 
With no model retraining, 𝗤𝗔𝗹𝗶𝗴𝗻 outperforms DPO-tuned models even when allowed to match inference compute, and achieves
Jiacheng Liu (@liujc1998) 's Twitter Profile Photo

As infini-gram surpasses 500 million API calls, today we're announcing two exciting updates: 1. Infini-gram is now open-source under Apache 2.0! 2. We indexed the training data of OLMo 2 models. Now you can search in the training data of these strong, fully-open LLMs. 🧵 (1/4)

Jiacheng Liu (@liujc1998) 's Twitter Profile Photo

Today we're unveiling OLMoTrace, a tool that enables everyone to understand the outputs of LLMs by connecting to their training data. We do this on unprecedented scale and in real time: finding matching text between model outputs and 4 trillion training tokens within seconds. ✨

Ai2 (@allen_ai) 's Twitter Profile Photo

"OLMoTrace is a breakthrough in AI development, setting a new standard for transparency and trust. We hope it will empower researchers, developers, and users to build with confidence—on models they can understand and trust." - CEO Ali Farhadi at tonight's fireside chat with

"OLMoTrace is a breakthrough in AI development, setting a new standard for transparency and trust. We hope it will empower researchers, developers, and users to build with confidence—on models they can understand and trust." - CEO Ali Farhadi at tonight's fireside chat with
Ian Magnusson (@ianmagnusson) 's Twitter Profile Photo

🔭 Science relies on shared artifacts collected for the common good. 🛰 So we asked: what's missing in open language modeling? 🪐 DataDecide 🌌 charts the cosmos of pretraining—across scales and corpora—at a resolution beyond any public suite of models that has come before.

Ximing Lu (@gximing) 's Twitter Profile Photo

With the rise of R1, search seems out of fashion? We prove the opposite! 😎 Introducing Retro-Search 🌈: an MCTS-inspired search algorithm that RETROspectively revises R1’s reasoning traces to synthesize untaken, new reasoning paths that are better 💡, yet shorter in length ⚡️.

With the rise of R1, search seems out of fashion? We prove the opposite! 😎

Introducing Retro-Search 🌈: an MCTS-inspired search algorithm that RETROspectively revises R1’s reasoning traces to synthesize untaken, new reasoning paths that are better 💡, yet shorter in length ⚡️.
Peter West (@peterwesttm) 's Twitter Profile Photo

I’ve been fascinated lately by the question: what kinds of capabilities might base LLMs lose when they are aligned? i.e. where can alignment make models WORSE? I’ve been looking into this with Christopher Potts and here's one piece of the answer: randomness and creativity

I’ve been fascinated lately by the question: what kinds of capabilities might base LLMs lose when they are aligned? i.e. where can alignment make models WORSE? I’ve been looking into this with <a href="/ChrisGPotts/">Christopher Potts</a> and here's one piece of the answer: randomness and creativity
Ai2 (@allen_ai) 's Twitter Profile Photo

📢We’re taking your questions now on Reddit for tomorrow’s AMA! Ask us anything about OLMo, our family of fully-open language models. Our researchers will be on hand to answer them Thursday, May 8 at 8am PST.

📢We’re taking your questions now on Reddit for tomorrow’s AMA! 

Ask us anything about OLMo, our family of fully-open language models. Our researchers will be on hand to answer them Thursday, May 8 at 8am PST.
Wenting Zhao (@wzhao_nlp) 's Twitter Profile Photo

Some personal news: I'll join UMass Amherst CS as an assistant professor in fall 2026. Until then, I'll postdoc at Meta nyc. Reasoning will continue to be my main interest, with a focus on data-centric approaches🤩 If you're also interested, apply to me (phds & a postdoc)!

Hyunwoo Kim (@hyunw_kim) 's Twitter Profile Photo

📢I'm thrilled to announce that I’ll be joining @KAIST_AI as an Assistant Professor in 2026, leading the Computation & Cognition (COCO) Lab🤖🧠: coco-kaist.github.io We'll be exploring reasoning, learning w/ synthetic data, and social agents! +I'm spending a gap year NVIDIA

📢I'm thrilled to announce that I’ll be joining @KAIST_AI as an Assistant Professor in 2026, leading the Computation &amp; Cognition (COCO) Lab🤖🧠: coco-kaist.github.io
We'll be exploring reasoning, learning w/ synthetic data, and social agents!
+I'm spending a gap year <a href="/nvidia/">NVIDIA</a>✨