Valentina Pyatkin (@valentina__py) 's Twitter Profile
Valentina Pyatkin

@valentina__py

Postdoc at the Allen Institute for AI @allen_ai and @uwnlp | on the academic job market

ID: 786941656859811840

linkhttps://valentinapy.github.io calendar_today14-10-2016 14:48:06

636 Tweet

2,2K Followers

1,1K Following

Yuling Gu (@gu_yuling) 's Twitter Profile Photo

Excited to be at #NAACL2025 in Albuquerque this week! I'll be presenting "OLMES: A Standard for Language Model Evaluations" (arxiv.org/abs/2406.08446)! Work done with my wonderful collaborators at Ai2 ❤️

Fazl Barez (@fazlbarez) 's Twitter Profile Photo

Responsible Reviewing #NeurIPS2025 — TL;DR 1- If you/ your co-author skip your assigned reviews → you wont see your own paper’s reviews. 2- Submit a poor quality review → your paper may be desk‑rejected. 👏 Nice one, NeurIPS! 🔗 blog.neurips.cc/2025/05/02/res…

Afra Amini (@afra_amini) 's Twitter Profile Photo

Current KL estimation practices in RLHF can generate high variance and even negative values! We propose a provably better estimator that only takes a few lines of code to implement.🧵👇 w/ Tim Vieira and Ryan Cotterell code: arxiv.org/pdf/2504.10637 paper: github.com/rycolab/kl-rb

Current KL estimation practices in RLHF can generate high variance and even negative values! We propose a provably better estimator that only takes a few lines of code to implement.🧵👇
w/ <a href="/xtimv/">Tim Vieira</a> and Ryan Cotterell
code: arxiv.org/pdf/2504.10637
paper: github.com/rycolab/kl-rb
Yanai Elazar (@yanaiela) 's Twitter Profile Photo

Lucas Beyer (bl16) (((ل()(ل() 'yoav))))👾 rohan anil arxiv.org/abs/2410.15002 This is in the text-to-image domain, and we have some ideas on how to extend this to the text domain We also recently published this: arxiv.org/abs/2504.12459, which connects the number of entities co-occurrences, and "linearity" in model representations:

Philippe Laban (@philippelaban) 's Twitter Profile Photo

🆕paper: LLMs Get Lost in Multi-Turn Conversation In real life, people don’t speak in perfect prompts. So we simulate multi-turn conversations — less lab-like, more like real use. We find that LLMs get lost in conversation. 👀What does that mean? 🧵1/N 📄arxiv.org/abs/2505.06120

🆕paper: LLMs Get Lost in Multi-Turn Conversation

In real life, people don’t speak in perfect prompts.
So we simulate multi-turn conversations — less lab-like, more like real use.

We find that LLMs get lost in conversation.
👀What does that mean? 🧵1/N
📄arxiv.org/abs/2505.06120
Jing-Jing Li (@drjingjing2026) 's Twitter Profile Photo

Excited to share that our SafetyAnalyst paper has been accepted to #icml2025! SafetyAnalyst provides a novel way to determine if some AI behavior would be safe. It’s accurate, interpretable, transparent, and steerable. 1/7

Excited to share that our SafetyAnalyst paper has been accepted to #icml2025! 

SafetyAnalyst provides a novel way to determine if some AI behavior would be safe. It’s accurate, interpretable, transparent, and steerable. 1/7
Ai2 (@allen_ai) 's Twitter Profile Photo

RewardBench 2 is here! We took a long time to learn from our first reward model evaluation tool to make one that is substantially harder and more correlated with both downstream RLHF and inference-time scaling.

RewardBench 2 is here! We took a long time to learn from our first reward model evaluation tool to make one that is substantially harder and more correlated with both downstream RLHF and inference-time scaling.
Sahil Verma (@sahil1v) 's Twitter Profile Photo

🚨 New Paper! 🚨 Guard models slow, language-specific, and modality-limited? Meet OmniGuard that detects harmful prompts across multiple languages & modalities all using one approach with SOTA performance in all 3 modalities!! while being 120X faster 🚀 arxiv.org/abs/2505.23856

🚨 New Paper! 🚨
Guard models slow, language-specific, and modality-limited?

Meet OmniGuard that detects harmful prompts across multiple languages &amp; modalities all using one approach with SOTA performance in all 3 modalities!! while being 120X faster 🚀

arxiv.org/abs/2505.23856
Saumya Malik (@saumyamalik44) 's Twitter Profile Photo

I’m thrilled to share RewardBench 2 📊— We created a new multi-domain reward model evaluation that is substantially harder than RewardBench, we trained and released 70 reward models, and we gained insights about reward modeling benchmarks and downstream performance!

I’m thrilled to share RewardBench 2 📊— We created a new multi-domain reward model evaluation that is substantially harder than RewardBench, we trained and released 70 reward models, and we gained insights about reward modeling benchmarks and downstream performance!
Eran Hirsch (@hirscheran) 's Twitter Profile Photo

🚨 Introducing LAQuer, accepted to #ACL2025 (main conf)! LAQuer provides more granular attribution for LLM generations: users can just highlight any output fact (top), and get attribution for that input snippet (bottom). This reduces the amount of text the user has to read by 2

🚨 Introducing LAQuer, accepted to #ACL2025 (main conf)!

LAQuer provides more granular attribution for LLM generations: users can just highlight any output fact (top), and get attribution for that input snippet (bottom). This reduces the amount of text the user has to read by 2
Jiacheng Liu (@liujc1998) 's Twitter Profile Photo

We enabled OLMoTrace for Tülu 3 models! 🤠 Matched spans are shorter than for OLMo models, bc we can only search in Tülu's post-training data (base model is Llama). Yet we thought it'd still bring some value. Try yourself on the Ai2 playground -- playground.allenai.org

We enabled OLMoTrace for Tülu 3 models! 🤠

Matched spans are shorter than for OLMo models, bc we can only search in Tülu's post-training data (base model is Llama). Yet we thought it'd still bring some value.

Try yourself on the Ai2 playground -- playground.allenai.org