Sweta Agrawal (@swetaagrawal20) 's Twitter Profile
Sweta Agrawal

@swetaagrawal20

Research Scientist @Google Translate |
Past: Postdoc Researcher @itnewspt | Ph.D. @ClipUmd, @umdcs
#nlproc

ID: 2559288140

linkhttp://sweta20.github.io calendar_today10-06-2014 15:47:30

339 Tweet

1,1K Followers

1,1K Following

Antonio Farinhas (@tozefarinhas) 's Twitter Profile Photo

If you're in Vancouver attending #NeurIPS2024, stop by our spotlight poster on Thu 12 Dec 11am-2pm PST (East Exhibit Hall A-C #3203). Check out the updated version of the paper: openreview.net/forum?id=rhCgiโ€ฆ

Artidoro Pagnoni (@artidoropagnoni) 's Twitter Profile Photo

๐Ÿš€ Introducing the Byte Latent Transformer (BLT) โ€“ An LLM architecture that scales better than Llama 3 using byte-patches instead of tokens ๐Ÿคฏ Paper ๐Ÿ“„ dl.fbaipublicfiles.com/blt/BLT__Patchโ€ฆ Code ๐Ÿ› ๏ธ github.com/facebookresearโ€ฆ

๐Ÿš€ Introducing the Byte Latent Transformer (BLT) โ€“ An LLM architecture that scales better than Llama 3 using byte-patches instead of tokens ๐Ÿคฏ 

Paper ๐Ÿ“„ dl.fbaipublicfiles.com/blt/BLT__Patchโ€ฆ
Code ๐Ÿ› ๏ธ github.com/facebookresearโ€ฆ
Saul Santos (@saul_santos1997) 's Twitter Profile Photo

๐Ÿš€ New paper alert! ๐Ÿš€ Ever tried asking an AI about a 2-hour movie? Yeahโ€ฆ not great. Check: โˆž-Video: A Training-Free Approach to Long Video Understanding via Continuous-Time Memory Consolidation! ๐Ÿ”— arxiv.org/abs/2501.19098 w/ Antonio Farinhas , Dan McNamee , Andre Martins

๐Ÿš€ New paper alert! ๐Ÿš€

Ever tried asking an AI about a 2-hour movie? Yeahโ€ฆ not great.

Check: โˆž-Video: A Training-Free Approach to Long Video Understanding via Continuous-Time Memory Consolidation! 

๐Ÿ”— arxiv.org/abs/2501.19098 
w/ <a href="/tozefarinhas/">Antonio Farinhas</a> , <a href="/mcneural_/">Dan McNamee</a> , <a href="/andre_t_martins/">Andre Martins</a>
Patrick Fernandes (@psanfernandes) 's Twitter Profile Photo

MT metrics excel at evaluating sentence translations, but struggle with complex texts We introduce *TREQA* a framework to assess how translations preserve key info by using LLMs to generate & answer questions about them arxiv.org/abs/2504.07583 (co-lead Sweta Agrawal) 1/15

MT metrics excel at evaluating sentence translations, but struggle with complex texts

We  introduce *TREQA* a framework to assess how translations preserve key info by using LLMs to generate &amp; answer questions about them

arxiv.org/abs/2504.07583

(co-lead <a href="/swetaagrawal20/">Sweta Agrawal</a>)

1/15
Mohit Iyyer (@mohitiyyer) 's Twitter Profile Photo

GRPO + BLEU is a surprisingly good combination for improving instruction following in LLMs, yielding results on par with those from strong reward models in our experiments! Check out our paper for more ๐Ÿ‘‡

Xin Eric Wang @ ICLR 2025 (@xwang_lk) 's Twitter Profile Photo

๐˜๐˜ถ๐˜ฎ๐˜ข๐˜ฏ๐˜ด ๐˜ต๐˜ฉ๐˜ช๐˜ฏ๐˜ฌ ๐˜ง๐˜ญ๐˜ถ๐˜ช๐˜ฅ๐˜ญ๐˜บโ€”๐˜ฏ๐˜ข๐˜ท๐˜ช๐˜จ๐˜ข๐˜ต๐˜ช๐˜ฏ๐˜จ ๐˜ข๐˜ฃ๐˜ด๐˜ต๐˜ณ๐˜ข๐˜ค๐˜ต ๐˜ค๐˜ฐ๐˜ฏ๐˜ค๐˜ฆ๐˜ฑ๐˜ต๐˜ด ๐˜ฆ๐˜ง๐˜ง๐˜ฐ๐˜ณ๐˜ต๐˜ญ๐˜ฆ๐˜ด๐˜ด๐˜ญ๐˜บ, ๐˜ง๐˜ณ๐˜ฆ๐˜ฆ ๐˜ง๐˜ณ๐˜ฐ๐˜ฎ ๐˜ณ๐˜ช๐˜จ๐˜ช๐˜ฅ ๐˜ญ๐˜ช๐˜ฏ๐˜จ๐˜ถ๐˜ช๐˜ด๐˜ต๐˜ช๐˜ค ๐˜ฃ๐˜ฐ๐˜ถ๐˜ฏ๐˜ฅ๐˜ข๐˜ณ๐˜ช๐˜ฆ๐˜ด. But current reasoning models remain constrained by discrete tokens, limiting their full

๐˜๐˜ถ๐˜ฎ๐˜ข๐˜ฏ๐˜ด ๐˜ต๐˜ฉ๐˜ช๐˜ฏ๐˜ฌ ๐˜ง๐˜ญ๐˜ถ๐˜ช๐˜ฅ๐˜ญ๐˜บโ€”๐˜ฏ๐˜ข๐˜ท๐˜ช๐˜จ๐˜ข๐˜ต๐˜ช๐˜ฏ๐˜จ ๐˜ข๐˜ฃ๐˜ด๐˜ต๐˜ณ๐˜ข๐˜ค๐˜ต ๐˜ค๐˜ฐ๐˜ฏ๐˜ค๐˜ฆ๐˜ฑ๐˜ต๐˜ด ๐˜ฆ๐˜ง๐˜ง๐˜ฐ๐˜ณ๐˜ต๐˜ญ๐˜ฆ๐˜ด๐˜ด๐˜ญ๐˜บ, ๐˜ง๐˜ณ๐˜ฆ๐˜ฆ ๐˜ง๐˜ณ๐˜ฐ๐˜ฎ ๐˜ณ๐˜ช๐˜จ๐˜ช๐˜ฅ ๐˜ญ๐˜ช๐˜ฏ๐˜จ๐˜ถ๐˜ช๐˜ด๐˜ต๐˜ช๐˜ค ๐˜ฃ๐˜ฐ๐˜ถ๐˜ฏ๐˜ฅ๐˜ข๐˜ณ๐˜ช๐˜ฆ๐˜ด. But current reasoning models remain constrained by discrete tokens, limiting their full
Edoardo Ponti (@pontiedoardo) 's Twitter Profile Photo

๐Ÿš€ By *learning* to compress the KV cache in Transformer LLMs, we can generate more tokens for the same compute budget. This unlocks *inference-time hyper-scaling* For the same runtime or memory load, we can boost LLM accuracy by pushing reasoning even further!

๐Ÿš€ By *learning* to compress the KV cache in Transformer LLMs, we can generate more tokens for the same compute budget. 

This unlocks *inference-time hyper-scaling*

For the same runtime or memory load, we can boost LLM accuracy by pushing reasoning even further!
ACLRollingReview (@reviewacl) 's Twitter Profile Photo

Dear ACL community, We are seeking emergency reviewers for the May cycle. Please indicate your availability (ASAP) if you can help review extra papers urgently (by the 24th of June AOE). Many thanks!

Manos Zaranis (@manoszaranis) 's Twitter Profile Photo

๐ŸšจMeet MFยฒ: Movieโ€ฏFacts &โ€ฏFibs: a new benchmark for long-movie understanding! ๐Ÿค”Do you think your model understands movies? Unlike existing benchmarks, MFยฒ targets memorable events, emotional arcs ๐Ÿ’”, and causal chains ๐Ÿ”— โ€” things humans recall easily, but even top models like

๐ŸšจMeet MFยฒ: Movieโ€ฏFacts &amp;โ€ฏFibs: a new benchmark for long-movie understanding!
๐Ÿค”Do you think your model understands movies?

Unlike existing benchmarks, MFยฒ targets memorable events, emotional arcs ๐Ÿ’”, and causal chains ๐Ÿ”— โ€” things humans recall easily, but even top models like
Guilherme Penedo (@gui_penedo) 's Twitter Profile Photo

We have finally released the ๐Ÿ“paper for ๐Ÿฅ‚FineWeb2, our large multilingual pre-training dataset. Along with general (and exhaustive) multilingual work, we introduce a concept that can also improve English performance: deduplication-based upsampling, which we call rehydration.

We have finally released the ๐Ÿ“paper for ๐Ÿฅ‚FineWeb2, our large multilingual pre-training dataset.

Along with general (and exhaustive) multilingual work, we introduce a concept that can also improve English performance: deduplication-based upsampling, which we call rehydration.
Olga Golovneva (@olganlp) 's Twitter Profile Photo

โœจMTA was accepted at #COLM2025 โœจ Since our first announcement, we have updated the paper with scaling laws, new baselines, and more evaluations! Code is now available in our repo: github.com/facebookresearโ€ฆ Conference on Language Modeling

ACL 2025 (@aclmeeting) 's Twitter Profile Photo

๐ŸŽ‰A reminder from ACL 2025: ๐Ÿ—ฃ๏ธ #InvitedTalk by Professor Luke Zettlemoyer. He'll be presenting on "Rethinking Pretraining: Data and Architecture." dive into the foundations of large language models! #ACL2025NLP #NLProc 2025.aclweb.org/program/keynotโ€ฆ

Google India (@googleindia) 's Twitter Profile Photo

If youโ€™re a student in India - youโ€™ve just been granted access to a FREE Gemini upgrade worth โ‚น19,500 for one year ๐Ÿฅณโœจ Claim and get free access to Veo 3, Gemini in Google apps, and 2TB storage ๐Ÿ”— goo.gle/freepro. Google Gemini App

Google DeepMind (@googledeepmind) 's Twitter Profile Photo

An advanced version of Gemini with Deep Think has officially achieved gold medal-level performance at the International Mathematical Olympiad. ๐Ÿฅ‡ It solved 5๏ธโƒฃ out of 6๏ธโƒฃ exceptionally difficult problems, involving algebra, combinatorics, geometry and number theory. Hereโ€™s how ๐Ÿงต

An advanced version of Gemini with Deep Think has officially achieved gold medal-level performance at the International Mathematical Olympiad. ๐Ÿฅ‡

It solved 5๏ธโƒฃ out of 6๏ธโƒฃ exceptionally difficult problems, involving algebra, combinatorics, geometry and number theory. Hereโ€™s how ๐Ÿงต
Diptesh Kanojia (@diptesh) 's Twitter Profile Photo

๐Ÿ“ข Test Set RELEASED! ๐Ÿš€ The test set for the #WMT25 Shared Task on QE-informed Segment-level Error Correction is now LIVE! It's time to put your MT error correction / APE methods to the test. Let's see how well they can correct machine translation! #NLProc #MT #WMT2025

Vilรฉm Zouhar (@zouharvi) 's Twitter Profile Photo

The 2025 MT Evaluation shared task brings together the strengths of the previous Metrics and Quality Estimation tasks under a single, unified evaluation framework. The following tasks are now open (deadline July 31st but participation has never been easier ๐Ÿ™‚)

Markus Freitag (@markuseful) 's Twitter Profile Photo

Our Google Translate team is bringing a strong presence to #ACL2025 in Vienna this week! ๐Ÿ‡ฆ๐Ÿ‡น My group is excited to present several of our latest papers. ๐Ÿ‘‡ Don't miss them!

Sweta Agrawal (@swetaagrawal20) 's Twitter Profile Photo

๐Ÿ“ขShared task deadline extended: You now have a whole week to go (until August 6 AoE) to register and send us your submissions!!

Sundar Pichai (@sundarpichai) 's Twitter Profile Photo

Excited to make our best AI tools free for college students in the US + other select countries for a year - and to provide $1B in funding for education + research, including free AI and career training for every college student in America.

Excited to make our best AI tools free for college students in the US + other select countries for a year - and to provide $1B in funding for education + research, including free AI and career training for every college student in America.