BigScience Research Workshop (@bigsciencew) 's Twitter Profile
BigScience Research Workshop

@bigsciencew

A research workshop on large language model gathering 1000+ researchers around the world

Follow the training of the 176B multilingual model live @BigScienceLLM

ID: 1385228159604465667

linkhttp://bigscience.huggingface.co calendar_today22-04-2021 13:45:29

354 Tweet

14,14K Followers

1 Following

BigCode (@bigcodeproject) 's Twitter Profile Photo

print("Hello world! 🎉") Excited to announce the BigCode project led by ServiceNow Research and Hugging Face! In the spirit of BigScience we aim to develop large language models for code in an open and responsible way. Join here: bigcode-project.org/docs/about/joi… A thread with our goals🧵

print("Hello world! 🎉")

Excited to announce the BigCode project led by <a href="/ServiceNowRSRCH/">ServiceNow Research</a> and <a href="/huggingface/">Hugging Face</a>! In the spirit of BigScience we aim to develop large language models for code in an open and responsible way.

Join here: bigcode-project.org/docs/about/joi…

A thread with our goals🧵
Sasha Luccioni, PhD 🦋🌎✨🤗 (@sashamtl) 's Twitter Profile Photo

The BigScience Research Workshop carbon footprint paper is live!! 🎉 Check it out to see how we calculated BLOOM's carbon footprint, covering all steps from the manufacturing of equipment 💻 to deployment! 🚀 arxiv.org/abs/2211.02001

Niklas Muennighoff (@muennighoff) 's Twitter Profile Photo

Crosslingual Generalization through Multitask Finetuning 🌸 Demo: huggingface.co/bigscience/blo… 📜 arxiv.org/abs/2211.01786 💻github.com/bigscience-wor… We present BLOOMZ & mT0, a family of models w/ up to 176B params that follow human instructions in >100 languages zero-shot. 1/7

Crosslingual Generalization through Multitask Finetuning  🌸

Demo: huggingface.co/bigscience/blo…
📜 arxiv.org/abs/2211.01786
💻github.com/bigscience-wor…

We present BLOOMZ &amp; mT0, a family of models w/ up to 176B params that follow human instructions in &gt;100 languages zero-shot. 1/7
BigScience Research Workshop (@bigsciencew) 's Twitter Profile Photo

Big day today with two papers out! BLOOM carbon footprint at arxiv.org/abs/2211.02001, new models BLOOMZ and mt0 at huggingface.co/bigscience/blo…

clem 🤗 (@clementdelangue) 's Twitter Profile Photo

The Bloom paper is out. Looks like it's doing worse than current GPT3 API in zero-shot generation tasks in English but better than other open-source LLMs & better than all in zs multi-lingual (which was the main goal). Proud of the work from the community! arxiv.org/abs/2211.05100

The Bloom paper is out. Looks like it's doing worse than current GPT3 API in zero-shot generation tasks in English but better than other open-source LLMs &amp; better than all in zs multi-lingual (which was the main goal). Proud of the work from the community! arxiv.org/abs/2211.05100
Max Ryabinin (@m_ryabinin) 's Twitter Profile Photo

Petals, a system for easy decentralized inference and adaptation of 100B+ LLMs, is now online! 🌸Generate text with BLOOM-176B using Colab or a desktop GPU 🔌Fine-tune large models for your tasks 👥Help others by contributing your GPUs or host a new swarm colab.research.google.com/drive/1Ervk6HP…

Petals, a system for easy decentralized inference and adaptation of 100B+ LLMs, is now online!

🌸Generate text with BLOOM-176B using Colab or a desktop GPU
🔌Fine-tune large models for your tasks
👥Help others by contributing your GPUs or host a new swarm
colab.research.google.com/drive/1Ervk6HP…
Yong Zheng-Xin (Yong) (@yong_zhengxin) 's Twitter Profile Photo

(Repost for corrected Arxiv) 🧐What’s the best way to quickly adapt large multilingual language models to new languages? We present our new paper from BigScience Research Workshop 🌸: BLOOM+1: Adding Language Support to BLOOM for Zero-Shot Prompting. 📜 arxiv.org/abs/2212.09535 [1/9]

(Repost for corrected Arxiv)
🧐What’s the best way to quickly adapt large multilingual language models to new languages? 

We present our new paper from <a href="/BigscienceW/">BigScience Research Workshop</a>  🌸:
BLOOM+1: Adding Language Support to BLOOM for Zero-Shot Prompting.

📜 arxiv.org/abs/2212.09535

[1/9]
Anna Rogers (@annargrs) 's Twitter Profile Photo

Worried about benchmark data contamination? Studying LLM memorization or attribution? BigScience Research Workshop BLOOM 🌸 now has exact & fuzzy search over full training data! with Ola Piktus🏆 Christopher Akiki Paulo Villegas Hugo Laurençon @ggdupont Sasha Luccioni, PhD 🦋🌎✨🤗 Yacine Jernite arxiv.org/abs/2302.14035 /1

Aran Komatsuzaki (@arankomatsuzaki) 's Twitter Profile Photo

The BigScience ROOTS Corpus: A 1.6TB Composite Multilingual Dataset Documents the data creation and curation efforts of ROOTS corpus, a 1.6TB dataset used to train BLOOM Releases a large initial subset of the corpus data: huggingface.co/bigscience-data abs: arxiv.org/abs/2303.03915

The BigScience ROOTS Corpus: A 1.6TB Composite Multilingual Dataset

Documents the data creation and curation efforts of ROOTS corpus, a 1.6TB dataset used to train BLOOM

Releases a large initial subset of the corpus

data: huggingface.co/bigscience-data
abs: arxiv.org/abs/2303.03915
Giada Pistilli (@giadapistilli) 's Twitter Profile Photo

As you already know, I am very proud of the collective work that enabled the development of BigScience Research Workshop's ethical charter. Today I am even more proud to announce that it's part of OECD Innovation's catalog to promote Trustworthy AI: such a milestone! oecd.ai/en/catalogue/t…

BigCode (@bigcodeproject) 's Twitter Profile Photo

Join us tomorrow, Wednesday 22nd (6:30 PM - 8:00PM CET) at the Mozilla Festival Science Fair to learn more about our work in the open and responsible development of large language models (LLMs) for code. schedule.mozillafestival.org/session/TJRU3L… #Mozfest

BigCode (@bigcodeproject) 's Twitter Profile Photo

Introducing: 💫StarCoder StarCoder is a 15B LLM for code with 8k context and trained only on permissive data in 80+ programming languages. It can be prompted to reach 40% pass@1 on HumanEval and act as a Tech Assistant. Try it here: shorturl.at/cYZ06r Release thread🧵

Introducing: 💫StarCoder

StarCoder is a 15B LLM for code with 8k context and trained only on permissive data in 80+ programming languages. It can be prompted to reach 40% pass@1 on HumanEval and act as a Tech Assistant.

Try it here: shorturl.at/cYZ06r

Release thread🧵
MMitchell (@mmitchell_ai) 's Twitter Profile Photo

If you wanted to see the fun panel/Q&A we did with Londoners on AI, you can check out the recording here! My preso at the start is also on Open Science, representing Hugging Face & BigScience Research Workshop.

Sasha Luccioni, PhD 🦋🌎✨🤗 (@sashamtl) 's Twitter Profile Photo

Never thought I'd see the day I'd have a publication in JMLR 🥹 So happy that the BLOOM carbon footprint paper has finally found a home at such an incredible venue! Thank you Shakir Mohamed for being such a great editor, it warms my heart to see your name on this paper 💚

Never thought I'd see the day I'd have a publication in JMLR 🥹 
So happy that the BLOOM carbon footprint paper has finally found a home at such an incredible venue!
Thank you <a href="/shakir_za/">Shakir Mohamed</a> for being such a great editor, it warms my heart to see your name on this paper 💚
Yacine Jernite (@yjernite) 's Twitter Profile Photo

I respect the caution, but also need to stress that efforts that pursue transparency as an operational value in service of actual inclusion and accountability do exist - see for example the writing on this very topic by BigScience Research Workshop, including its ethical charter. 1/3

Stas Bekman (@stasbekman) 's Twitter Profile Photo

The Universal Checkpointing paper is out! arxiv.org/abs/2406.18820 If you remember the BigScience Research Workshop BLOOM-176B training, Tunji Ruwase and I co-invented this technology for Megatron-Deepspeed in order to enable to quickly scale up and down node topology while continuing training.

Oxford Internet Institute (@oiioxford) 's Twitter Profile Photo

DPhil candidate @cailean_osborne shares reflections on the Open Source Initiative @[email protected] co-design process to define #opensourceAI and recommends next steps, including improving model safety and supporting more grassroots initiatives like BigScience Research Workshop.