Thibault Sellam (@thiboibo) 's Twitter Profile
Thibault Sellam

@thiboibo

ID: 149132642

linkhttp://sellam.me calendar_today28-05-2010 14:05:36

45 Tweet

342 Followers

276 Following

Sebastian Gehrmann (@sebgehr) 's Twitter Profile Photo

Listing issues in NLG evaluations turned into a 25 page survey! In “Repairing the Cracked Foundation: A Survey of Obstacles in Evaluation Practices for Generated Text”, Thibault Sellam Elizabeth Clark and I cover 250+ papers. 📄Link: arxiv.org/abs/2202.06935 Want to learn more?👇

Listing issues in NLG evaluations turned into a 25 page survey!

In “Repairing the Cracked Foundation: A Survey of Obstacles in Evaluation Practices for Generated Text”, <a href="/ThiboIbo/">Thibault Sellam</a> <a href="/eaclark07/">Elizabeth Clark</a> and I cover 250+ papers. 
📄Link: arxiv.org/abs/2202.06935 

Want to learn more?👇
Dipanjan Das (@dipanjand) 's Twitter Profile Photo

My colleagues Chris Alberti, Kuzman Ganchev and I are looking for a fall intern. The topic is improving language technologies for underrepresented languages by leveraging large pretrained language models. The position could be in NYC or remote. If interested, please reach out.

Kelvin Guu (@kelvin_guu) 's Twitter Profile Photo

New from Google Research! Language models perform amazing feats, but often still "hallucinate" unsupported content. Our model, RARR🐯, automatically researches & revises the output of any LM to fix hallucinations and provide citations for each sentence. arxiv.org/abs/2210.08726 🧵

New from Google Research! Language models perform amazing feats, but often still "hallucinate" unsupported content. Our model, RARR🐯, automatically researches &amp; revises the output of any LM to fix hallucinations and provide citations for each sentence. arxiv.org/abs/2210.08726 🧵
Ran TIAN (@robin_tian) 's Twitter Profile Photo

Amos is a new optimizer that we propose to pre-train large language models. It is more efficient and converges faster than AdamW: ≤ 51% memory for slot variables, and better valid loss within ≤ 70% training time! New from Google Research. Preprint: arxiv.org/abs/2210.11693

Amos is a new optimizer that we propose to pre-train large language models. It is more efficient and converges faster than AdamW: ≤ 51% memory for slot variables, and better valid loss within ≤ 70% training time!

New from Google Research. Preprint: arxiv.org/abs/2210.11693
Elizabeth Clark (@eaclark07) 's Twitter Profile Photo

We are excited to release Seahorse 🌊🐴, a ✨multilingual, multifaceted summarization evaluation dataset✨ 96,000+ human ratings to enable faster progress in training and evaluating learnt metrics for summarization! Preprint: arxiv.org/abs/2305.13194 Data: goo.gle/seahorse

Google AI (@googleai) 's Twitter Profile Photo

Learn how SQuId (Speech Quality Identification), a 600M parameter regression model that describes to what extent a piece of speech sounds natural, can be used to complement human ratings for the text-to-speech evaluation of many languages → goo.gle/3J1kghg

Learn how SQuId (Speech Quality Identification), a 600M parameter regression model that describes to what extent a piece of speech sounds natural, can be used to complement human ratings for the text-to-speech evaluation of many languages → goo.gle/3J1kghg
Sam Fraiberger 🔎🌍 (@spfraib) 's Twitter Profile Photo

🚨 Dream Job Alert 🚨 We are looking to fill various positions including NLP researcher, data engineer and software developer. If you are interested in #LLM, causal inference and having a positive impact on the world, please reach out! worldbank.org/en/research/di…

Thibault Sellam (@thiboibo) 's Twitter Profile Photo

The Searhorse dataset is available - 96K ratings to train and evaluate new summarization metrics. Congrats Elizabeth Clark and team! Paper here: arxiv.org/abs/2305.13194 Models here: huggingface.co/collections/go…

Pete Shaw (@ptshaw2) 's Twitter Profile Photo

Excited to share new work from Google DeepMind: “ProtEx: A Retrieval-Augmented Approach for Protein Function Prediction” biorxiv.org/content/10.110…

Excited to share new work from <a href="/GoogleDeepMind/">Google DeepMind</a>: “ProtEx: A Retrieval-Augmented Approach for Protein Function Prediction”

biorxiv.org/content/10.110…
Aran Komatsuzaki (@arankomatsuzaki) 's Twitter Profile Photo

Google presents Can Long-Context Language Models Subsume Retrieval, RAG, SQL, and More? Long-context LM: - Often rivals SotA retrieval and RAG systems - But still struggles with areas like compositional reasoning repo: github.com/google-deepmin… abs: arxiv.org/abs/2406.13121

Google presents Can Long-Context Language Models Subsume Retrieval, RAG, SQL, and More?

Long-context LM:
- Often rivals SotA retrieval and RAG systems
- But still struggles with areas like compositional reasoning

repo: github.com/google-deepmin…
abs: arxiv.org/abs/2406.13121
iseeaswell꩜bʂky (@iseeaswell) 's Twitter Profile Photo

The playlist starts in West Africa and wanders West-to-East until it hits Brazil. It’s composed of songs supplied by native speakers, which tend to be bangers. youtube.com/playlist?list=…

Jacob Austin (@jacobaustin132) 's Twitter Profile Photo

Making LLMs run efficiently can feel scary, but scaling isn’t magic, it’s math! We wanted to demystify the “systems view” of LLMs and wrote a little textbook called “How To Scale Your Model” which we’re releasing today. 1/n

Making LLMs run efficiently can feel scary, but scaling isn’t magic, it’s math! We wanted to demystify the “systems view” of LLMs and wrote a little textbook called “How To Scale Your Model” which we’re releasing today. 1/n