Oreva Ahia (@orevaahia) 's Twitter Profile
Oreva Ahia

@orevaahia

PhD student @uwcse | ex: AI/ML Research Intern @apple | Co-organizer @AISaturdayLagos | Researcher @MasakhaneNLP --Tomorrow may never come !

ID: 836314434

linkhttps://orevaahia.github.io/ calendar_today20-09-2012 20:39:09

1,1K Tweet

1,1K Followers

1,1K Following

Valentin Hofmann (@vjhofmann) 's Twitter Profile Photo

Humans store thousands of multi-word expressions like "of course" in their mental lexicon, but current tokenizers don't support multi-word tokens. Enter SuperBPE, a tokenizer that lifts this restriction and brings substantial gains in efficiency and performance! ๐Ÿš€ Details ๐Ÿ‘‡

Gonรงalo Faria (@goncalorafaria) 's Twitter Profile Photo

Introducing ๐—ค๐—”๐—น๐—ถ๐—ด๐—ป๐Ÿš€, a ๐˜๐—ฒ๐˜€๐˜-๐˜๐—ถ๐—บ๐—ฒ ๐—ฎ๐—น๐—ถ๐—ด๐—ป๐—บ๐—ฒ๐—ป๐˜ ๐—บ๐—ฒ๐˜๐—ต๐—ผ๐—ฑ that improves language model performance using Markov chain Monte Carlo. With no model retraining, ๐—ค๐—”๐—น๐—ถ๐—ด๐—ป outperforms DPO-tuned models even when allowed to match inference compute, and achieves

Introducing ๐—ค๐—”๐—น๐—ถ๐—ด๐—ป๐Ÿš€, a ๐˜๐—ฒ๐˜€๐˜-๐˜๐—ถ๐—บ๐—ฒ ๐—ฎ๐—น๐—ถ๐—ด๐—ป๐—บ๐—ฒ๐—ป๐˜ ๐—บ๐—ฒ๐˜๐—ต๐—ผ๐—ฑ that improves language model performance using Markov chain Monte Carlo. 
With no model retraining, ๐—ค๐—”๐—น๐—ถ๐—ด๐—ป outperforms DPO-tuned models even when allowed to match inference compute, and achieves
Sachin Kumar (@shocheen) 's Twitter Profile Photo

Really excited for this paper to be out. This project began nearly a year ago when I was at Ai2. Activation steering and related ideas were incredibly appealing, and we explored applying them to a range of problems. But none of the techniques we tried led to meaningful

Cohere Labs (@cohere_labs) 's Twitter Profile Photo

We're starting a new chapter as Cohere Labs! ๐ŸŽ‰ After 3 years of innovation, our new name reflects our continued dedication to research, collaboration, and open science. Our mission remains: transforming spaces where breakthroughs happen. Hereโ€™s to the next chapter.

Sara Hooker (@sarahookr) 's Twitter Profile Photo

Very proud of what we have achieved over last 3 years, and the breakthroughs ahead. ๐Ÿ”ฅ Our new name helps better communicate our work and our impact at the frontier of AI progress. Everything else stays the same, including our commitment to explore the unknown, together.

Ian Magnusson (@ianmagnusson) 's Twitter Profile Photo

๐Ÿ”ญ Science relies on shared artifacts collected for the common good. ๐Ÿ›ฐ So we asked: what's missing in open language modeling? ๐Ÿช DataDecide ๐ŸŒŒ charts the cosmos of pretrainingโ€”across scales and corporaโ€”at a resolution beyond any public suite of models that has come before.

Julie Kallini โœจ @ ICLR 2025 โœˆ๏ธ (@juliekallini) 's Twitter Profile Photo

๐Ÿš€ In T-minus 1 week, Iโ€™ll be at ICLR presenting MrT5! The final version has tons of updates: - New controller algorithm for targeted compression rates - More baselines and downstream tasks - Scaled-up experiments to 1.23B parameter models And now, MrT5 is on ๐Ÿค—HuggingFace! ๐Ÿงต

๐Ÿš€ In T-minus 1 week, Iโ€™ll be at ICLR presenting MrT5!

The final version has tons of updates:
- New controller algorithm for targeted compression rates
- More baselines and downstream tasks
- Scaled-up experiments to 1.23B parameter models

And now, MrT5 is on ๐Ÿค—HuggingFace! ๐Ÿงต
Tokenization Workshop (TokShop) @ICML2025 (@tokshop2025) 's Twitter Profile Photo

๐Ÿšจ NEW WORKSHOP ALERT ๐Ÿšจ We're thrilled to announce the first-ever Tokenization Workshop (TokShop) at #ICML2025 ICML Conference! ๐ŸŽ‰ Submissions are open for work on tokenization across all areas of machine learning. ๐Ÿ“… Submission deadline: May 30, 2025 ๐Ÿ”— tokenization-workshop.github.io

Oreva Ahia (@orevaahia) 's Twitter Profile Photo

Working on tokenization across any modality, text, audio, images, videos ? Submit your paper to our Tokenization Workshop at #ICML2025!

Valentin Hofmann (@vjhofmann) 's Twitter Profile Photo

Delighted there will finally be a workshop devoted to tokenization - a critical topic for LLMs and beyond! ๐ŸŽ‰ Join us for the inaugural edition of TokShop at #ICML2025 ICML Conference in Vancouver this summer! ๐Ÿค—

Kabir (@kabirahuja004) 's Twitter Profile Photo

๐Ÿ“ข New Paper! Tired ๐Ÿ˜ด of reasoning benchmarks full of math & code? In our work we consider the problem of reasoning for plot holes in stories -- inconsistencies in a storyline that break the internal logic or rules of a storyโ€™s world ๐ŸŒŽ W/ Melanie Sclar, and tsvetshop 1/n

๐Ÿ“ข New Paper!

Tired ๐Ÿ˜ด of reasoning benchmarks full of math & code? In our work we consider the problem of reasoning for plot holes in stories -- inconsistencies in a storyline that break the internal logic or rules of a storyโ€™s world ๐ŸŒŽ

W/ <a href="/melaniesclar/">Melanie Sclar</a>, and <a href="/tsvetshop/">tsvetshop</a>

1/n
Chan Young Park (@chan_young_park) 's Twitter Profile Photo

While I'm on X to share my paper, I also have a life update I'll be joining School of Information - UT Austin as an assistant professor starting Fall 2026! Excited for this next chapter, and to keep working on teaching computers to better understand language and humans (+now teaching humans too)

Stella Li (@stellalisy) 's Twitter Profile Photo

๐Ÿคฏ We cracked RLVR with... Random Rewards?! Training Qwen2.5-Math-7B with our Spurious Rewards improved MATH-500 by: - Random rewards: +21% - Incorrect rewards: +25% - (FYI) Ground-truth rewards: + 28.8% How could this even workโ‰๏ธ Here's why: ๐Ÿงต Blogpost: tinyurl.com/spurious-rewarโ€ฆ

๐Ÿคฏ We cracked RLVR with... Random Rewards?!
Training Qwen2.5-Math-7B with our Spurious Rewards improved MATH-500 by:
- Random rewards: +21%
- Incorrect rewards: +25%
- (FYI) Ground-truth rewards: + 28.8%
How could this even workโ‰๏ธ Here's why: ๐Ÿงต
Blogpost: tinyurl.com/spurious-rewarโ€ฆ
Roy Xie (@royxie_) 's Twitter Profile Photo

Can we train reasoning LLMs to generate answers as they think? Introducing ๐ˆ๐ง๐ญ๐ž๐ซ๐ฅ๐ž๐š๐ฏ๐ž๐ ๐‘๐ž๐š๐ฌ๐จ๐ง๐ข๐ง๐ ! We train LLMs to alternate between thinking & answering ๐Ÿš€ Reducing Time-to-First-Token (TTFT) by over 80% โšกAND improving Pass@1 accuracy up to 19.3%!๐Ÿ“ˆ ๐Ÿงต 1/n

Can we train reasoning LLMs to generate answers as they think?
Introducing ๐ˆ๐ง๐ญ๐ž๐ซ๐ฅ๐ž๐š๐ฏ๐ž๐ ๐‘๐ž๐š๐ฌ๐จ๐ง๐ข๐ง๐ ! We train LLMs to  alternate between thinking &amp; answering ๐Ÿš€
Reducing Time-to-First-Token (TTFT) by over 80% โšกAND improving Pass@1 accuracy up to 19.3%!๐Ÿ“ˆ

๐Ÿงต 1/n
Tokenization Workshop (TokShop) @ICML2025 (@tokshop2025) 's Twitter Profile Photo

Got a good tokenization paper under review at COLM, but the scores were a letdown? ๐Ÿ˜ฌ Why bother with rebuttal when the perfect venue is right around the corner! Submit your paper to the #ICML2025 Tokenization Workshop (TokShop) by May 30! ๐Ÿš€

Yizhong Wang (@yizhongwyz) 's Twitter Profile Photo

Thrilled to announce that I will be joining UT Austin Computer Science at UT Austin as an assistant professor in fall 2026! I will continue working on language models, data challenges, learning paradigms, & AI for innovation. Looking forward to teaming up with new students & colleagues! ๐Ÿค ๐Ÿค˜

Thrilled to announce that I will be joining <a href="/UTAustin/">UT Austin</a> <a href="/UTCompSci/">Computer Science at UT Austin</a> as an assistant professor in fall 2026! 

I will continue working on language models, data challenges, learning paradigms, &amp; AI for innovation. Looking forward to teaming up with new students &amp; colleagues! ๐Ÿค ๐Ÿค˜
Oreva Ahia (@orevaahia) 's Twitter Profile Photo

๐Ÿšจ Reminder: Paper submissions for the 1st Tokenization Workshop (TokShop) at #ICML2025 are due today May 30! ๐Ÿ”—CFP: tokenization-workshop.github.io

Sara Hooker (@sarahookr) 's Twitter Profile Photo

Truly excellent video by Machine Learning Street Talk about how a handful of providers have systematically overfit to lmarena.ai. 26 mins of video showcase how easy it has been to distort the rankings. As scientists, we must do better. As a community, I hope we can demand better.

Truly excellent video by <a href="/MLStreetTalk/">Machine Learning Street Talk</a> about how a handful of providers have systematically overfit to <a href="/lmarena_ai/">lmarena.ai</a>.

26 mins of video showcase how easy it has been to distort the rankings. 

As scientists, we must do better. As a community, I hope we can demand better.