
BigScience Research Workshop
@bigsciencew
A research workshop on large language model gathering 1000+ researchers around the world
Follow the training of the 176B multilingual model live @BigScienceLLM
ID: 1385228159604465667
http://bigscience.huggingface.co 22-04-2021 13:45:29
354 Tweet
14,14K Followers
1 Following

print("Hello world! 🎉") Excited to announce the BigCode project led by ServiceNow Research and Hugging Face! In the spirit of BigScience we aim to develop large language models for code in an open and responsible way. Join here: bigcode-project.org/docs/about/joi… A thread with our goals🧵


The BigScience Research Workshop carbon footprint paper is live!! 🎉 Check it out to see how we calculated BLOOM's carbon footprint, covering all steps from the manufacturing of equipment 💻 to deployment! 🚀 arxiv.org/abs/2211.02001




(Repost for corrected Arxiv) 🧐What’s the best way to quickly adapt large multilingual language models to new languages? We present our new paper from BigScience Research Workshop 🌸: BLOOM+1: Adding Language Support to BLOOM for Zero-Shot Prompting. 📜 arxiv.org/abs/2212.09535 [1/9]
![Yong Zheng-Xin (Yong) (@yong_zhengxin) on Twitter photo (Repost for corrected Arxiv)
🧐What’s the best way to quickly adapt large multilingual language models to new languages?
We present our new paper from <a href="/BigscienceW/">BigScience Research Workshop</a> 🌸:
BLOOM+1: Adding Language Support to BLOOM for Zero-Shot Prompting.
📜 arxiv.org/abs/2212.09535
[1/9] (Repost for corrected Arxiv)
🧐What’s the best way to quickly adapt large multilingual language models to new languages?
We present our new paper from <a href="/BigscienceW/">BigScience Research Workshop</a> 🌸:
BLOOM+1: Adding Language Support to BLOOM for Zero-Shot Prompting.
📜 arxiv.org/abs/2212.09535
[1/9]](https://pbs.twimg.com/media/Fka03d2WIAA0Iz4.jpg)

Worried about benchmark data contamination? Studying LLM memorization or attribution? BigScience Research Workshop BLOOM 🌸 now has exact & fuzzy search over full training data! with Ola Piktus🏆 Christopher Akiki Paulo Villegas Hugo Laurençon @ggdupont Sasha Luccioni, PhD 🦋🌎✨🤗 Yacine Jernite arxiv.org/abs/2302.14035 /1


As you already know, I am very proud of the collective work that enabled the development of BigScience Research Workshop's ethical charter. Today I am even more proud to announce that it's part of OECD Innovation's catalog to promote Trustworthy AI: such a milestone! oecd.ai/en/catalogue/t…

Join us tomorrow, Wednesday 22nd (6:30 PM - 8:00PM CET) at the Mozilla Festival Science Fair to learn more about our work in the open and responsible development of large language models (LLMs) for code. schedule.mozillafestival.org/session/TJRU3L… #Mozfest


If you wanted to see the fun panel/Q&A we did with Londoners on AI, you can check out the recording here! My preso at the start is also on Open Science, representing Hugging Face & BigScience Research Workshop.

Never thought I'd see the day I'd have a publication in JMLR 🥹 So happy that the BLOOM carbon footprint paper has finally found a home at such an incredible venue! Thank you Shakir Mohamed for being such a great editor, it warms my heart to see your name on this paper 💚


I respect the caution, but also need to stress that efforts that pursue transparency as an operational value in service of actual inclusion and accountability do exist - see for example the writing on this very topic by BigScience Research Workshop, including its ethical charter. 1/3

The top 15 most-liked organizations on Hugging Face 1. Stability AI 20k likes 2. AI at Meta 20k 3. Runway 11k 4. CompVis 10k 5. Tsinghua KEG (THUDM) 7k 6. BigScience Research Workshop 7k 7. Technology Innovation Institute 7k 8. Microsoft 6.5k 9. Google AI 6k 10. OpenAI 4k 11. BigCode 4k 12. . 4k 13. UKP Lab 3k

The Universal Checkpointing paper is out! arxiv.org/abs/2406.18820 If you remember the BigScience Research Workshop BLOOM-176B training, Tunji Ruwase and I co-invented this technology for Megatron-Deepspeed in order to enable to quickly scale up and down node topology while continuing training.

DPhil candidate @cailean_osborne shares reflections on the Open Source Initiative @[email protected] co-design process to define #opensourceAI and recommends next steps, including improving model safety and supporting more grassroots initiatives like BigScience Research Workshop.
