Hanna Hajishirzi (@hannahajishirzi) 's Twitter Profile
Hanna Hajishirzi

@hannahajishirzi

Sr. Director at @allen_ai, Prof at @uw_cse; co-lead OLMo, Tulu; AI/NLP/ML researcher at @uw_nlp

ID: 1098819287446253569

linkhttps://homes.cs.washington.edu/~hannaneh/ calendar_today22-02-2019 05:38:27

661 Tweet

7,7K Followers

397 Following

Ai2 (@allen_ai) 's Twitter Profile Photo

We are #1 on the Hugging Face heatmap - this is what true openness looks like!🥇🎉 750+ models 230+ datasets And counting... Come build with us huggingface.co/spaces/cfahlgr…

Hanna Hajishirzi (@hannahajishirzi) 's Twitter Profile Photo

Yayyy!!! Best paper honorable mention at CVPR goes to our Molmo and Pixmo Ai2! This is now becoming a tend :) Last year both OLMo and Dolma received best paper awards at ACL.

Ai2 (@allen_ai) 's Twitter Profile Photo

New updates for olmOCR, our fully open toolkit for transforming documents (PDFs & images) into clean markdown. We released: 1️⃣ New benchmark for fair comparison of OCR engines and APIs 2️⃣ Improved inference that is faster and cheaper to run 3️⃣ Docker image for easy deployment

New updates for olmOCR, our fully open toolkit for transforming documents (PDFs & images) into clean markdown. We released:

1️⃣ New benchmark for fair comparison of OCR engines and APIs
2️⃣ Improved inference that is faster and cheaper to run
3️⃣ Docker image for easy deployment
Ai2 (@allen_ai) 's Twitter Profile Photo

How well can today’s models generalize, compose, or even innovate on unseen problems? OMEGA Ω is a new math benchmark that pushes LLMs beyond pattern-matching to test true mathematical reasoning. ⚡ allenai.org/blog/omega

Joongwon Kim (@danieljwkim) 's Twitter Profile Photo

Can we improve Llama 3’s reasoning abilities through post-training only? Introducing ASTRO, our new framework that teaches LLMs to perform in-context search and generate long CoT to solve math problems, via SFT and RL. Work done at @aiatmeta. 📄 Paper: arxiv.org/abs/2507.00417

Victoria Graf (@victoriawgraf) 's Twitter Profile Photo

Worried about overfitting to IFEval? 🤔 Use ✨IFBench✨ our new, challenging instruction-following benchmark! Loved working w/ Valentina Pyatkin! Personal highlight: our multi-turn eval setting makes it possible to isolate constraint-following from the rest of the instruction 🔍

Cirrascale Cloud Services (@cirrascale) 's Twitter Profile Photo

🚨 Just announced: OLMo, Molmo & Tülu are now LIVE on the Cirrascale Inference Platform! It’s official, Cirrascale is the first to offer commercial inference endpoints for @Ai2’s OLMo, Molmo & Tülu models on our Inference Platform. Our Inference Platform provides a fully open,

🚨 Just announced: OLMo, Molmo & Tülu are now LIVE on the Cirrascale Inference Platform!

It’s official, Cirrascale is the first to offer commercial inference endpoints for @Ai2’s OLMo, Molmo & Tülu models on our Inference Platform.

Our Inference Platform provides a fully open,
Zhiyuan Zeng (@zhiyuanzeng_) 's Twitter Profile Photo

EvalTree accepted to Conference on Language Modeling 2025 - my first PhD work and first COLM paper 🙌! What would you like to see next—extensions, applications, or other directions? Always open to ideas! 🧐

EvalTree accepted to <a href="/COLM_conf/">Conference on Language Modeling</a> 2025 - my first PhD work and first COLM paper 🙌!

What would you like to see next—extensions, applications, or other directions? Always open to ideas! 🧐
Weijia Shi (@weijiashi2) 's Twitter Profile Photo

Can data owners & LM developers collaborate to build a strong shared model while each retaining data control? Introducing FlexOlmo💪, a mixture-of-experts LM enabling: • Flexible training on your local data without sharing it • Flexible inference to opt in/out your data

Niklas Muennighoff (@muennighoff) 's Twitter Profile Photo

Sparse Mixture-of-Expert LLMs to opt data in & out on the fly — I think a compelling vision for a future where AI developers & publishers work together rather than filing lawsuits🙂

Sparse Mixture-of-Expert LLMs to opt data in &amp; out on the fly — I think a compelling vision for a future where AI developers &amp; publishers work together rather than filing lawsuits🙂
Sewon Min (@sewon__min) 's Twitter Profile Photo

It has been great working on the project with support from Ai2! I believe there are many meaningful ways different people and orgs can work together to build strong shared models, and data collaboration might be the most impactful form of it. 📄Paper:

Will Knight (@willknight) 's Twitter Profile Photo

New on WIRED: A novel type of distributed mixture-of-experts model from Ai2 (called FlexOlmo) allows data can be contributed to a frontier model confidentially, and even revoked after the model is built: wired.com/story/flexolmo…