Ola Piktus (@olapiktus) 's Twitter Profile
Ola Piktus

@olapiktus

ID: 1102704797403238400

calendar_today04-03-2019 22:58:05

70 Tweet

1,1K Followers

405 Following

WikiResearch (@wikiresearch) 's Twitter Profile Photo

"Improving Wikipedia Verifiability with AI " a system that identifies Wikipedia citations that are unlikely to support their claims, and subsequently recommend better ones from the web. (Petroni et al, 2022) openreview.net/forum?id=qfTqR…

"Improving Wikipedia Verifiability with AI " a system that identifies Wikipedia citations that are unlikely to support their claims, and subsequently recommend better ones from the web.

(Petroni et al, 2022)

openreview.net/forum?id=qfTqR…
Will Held (@williambarrheld) 's Twitter Profile Photo

The Roots Search Tool from Hugging Face allowed me to quickly find XNLI examples in the BLOOM pre-training data - probably true for other benchmarks too. This 🔥 tool really highlights open-data as key to making the study of LLM capabilities a science. huggingface.co/spaces/bigscie…

The Roots Search Tool from <a href="/huggingface/">Hugging Face</a> allowed me to quickly find XNLI examples in the BLOOM pre-training data - probably true for other benchmarks too. 

This 🔥 tool really highlights open-data as key to making the study of LLM capabilities a science.

huggingface.co/spaces/bigscie…
Anna Rogers (@annargrs) 's Twitter Profile Photo

📢 New blog post: the attribution problem with generative AI #NLProc AI & Society TLDR: Some argue that publicly available data is fair game for commercial models bc human text/art also has sources. But unlike models, we know when attribution is due... hackingsemantics.xyz/2022/attributi…

Ola Piktus (@olapiktus) 's Twitter Profile Photo

ROOTS Search Tool now shinier and better with both exact search and BM25-based sparse retrieval ☀️🌸 Check out Anna Rogers 's thread for details 🚀

Anna Rogers (@annargrs) 's Twitter Profile Photo

ROOTS search tool for BLOOM🌸 training data will be presented at #ACL2023 as a demo paper! Really proud to be part of this important precedent.

Niklas Muennighoff (@muennighoff) 's Twitter Profile Photo

How to keep scaling Large Language Models when data runs out? 🎢 We train 400 models with up to 9B params & 900B tokens to create an extension of Chinchilla scaling laws for repeated data. Results are interesting… 🧐 📜: arxiv.org/abs/2305.16264 1/7

How to keep scaling Large Language Models when data runs out? 🎢

We train 400 models with up to 9B params &amp; 900B tokens to create an extension of Chinchilla scaling laws for repeated data. Results are interesting… 🧐

📜:  arxiv.org/abs/2305.16264

1/7
EleutherAI (@aieleuther) 's Twitter Profile Photo

Releasing data is amazing, but tools like these that help people make sense of the data is arguably an even more important step forward for data transparency. We're thrilled to see our community continue to lead by example when it comes to in transparent releases.

Fabio Petroni (@fabio_petroni) 's Twitter Profile Photo

Every researcher's path is a rollercoaster filled with numerous lows. With unspeakable joy, today I want to share a high point: our paper, “Improving Wikipedia Verifiability with AI,” has been published in the prestigious Nature Portfolio. Dive in nature.com/articles/s4225… [1/4]

Every researcher's path is a rollercoaster filled with numerous lows. With unspeakable joy, today I want to share a high point: our paper, “Improving Wikipedia Verifiability with AI,” has been published in the prestigious <a href="/NaturePortfolio/">Nature Portfolio</a>. Dive in nature.com/articles/s4225… [1/4]
Ola Piktus (@olapiktus) 's Twitter Profile Photo

Last time in New Orleans was fun and it was not even for NeurIPS. If you wanna chat about RAG, god or data hit me up this week. If you wanna tell me about LLM UX please ping me too 🙌 #NeurIPS2023

Last time in New Orleans was fun and it was not even for NeurIPS. If you wanna chat about RAG, god or data hit me up this week. If you wanna tell me about LLM UX please ping me too 🙌 #NeurIPS2023
Patrick Lewis (@psh_lewis) 's Twitter Profile Photo

New paper from our team, led by @pat_verga Are you: * Doing evaluation with LLMs? * Using a huge model? * Worried about self-recognition? Try an ensemble of smaller LLMs. Use a PoLL: less biased, faster, 7x cheaper. Works great on QA & Arena-hard evals arxiv.org/abs/2404.18796

New paper from our team, led by @pat_verga 
Are you:
* Doing evaluation with LLMs? 
* Using a huge model?
* Worried about self-recognition?

Try an ensemble of smaller LLMs.
Use a PoLL: less biased, faster, 7x cheaper. Works great on QA &amp; Arena-hard evals
arxiv.org/abs/2404.18796