Bertrand Charpentier (@bertrand_charp) 's Twitter Profile
Bertrand Charpentier

@bertrand_charp

Founder, President & Chief Scientist @PrunaAI | Prev. @Twitter research, Ph.D. in ML @TU_Muenchen

@bertrand-sharp.bsky.social
@[email protected]

ID: 1080138653807058945

linkhttps://sharpenb.github.io/ calendar_today01-01-2019 16:28:17

281 Tweet

490 Followers

132 Following

David Berenstein (@davidberenstei) 's Twitter Profile Photo

๐Ÿ’ฅ SMASH and run models 5x faster, 5x cheaper Pruna AI is the AI Optimization Engine for ML teams seeking to simplify scalable inference. Make sure to โญ๏ธ their GitHub: buff.ly/4tE5Ahy TechCrunch: buff.ly/64jQYu3 Smashed Models on HF: buff.ly/iV3XDeU

Bertrand Charpentier (@bertrand_charp) 's Twitter Profile Photo

We open-sourced the pruna package! ๐ŸŒ๐Ÿš€ - It supports various compression methods such as pruning, quantization, distillation, and caching that can be combined! - It enables easy evaluation of efficiency and quality of compressed models! Now, developers around the world can

Pruna AI (@prunaai) 's Twitter Profile Photo

๐Ÿ’ฆ Everythingโ€™s fineโ€ฆ until it isnโ€™t. ๐Ÿ”ฅ Generative AI runs on code and concrete, metals, heat, and vapor. Letโ€™s talk about the real infrastructure behind your prompts. ๐Ÿ—โš™๏ธ๐Ÿ”ฅ๐Ÿ’จ ๐Ÿ—“ April 15, 5PM UTC ๐Ÿ”— linkedin.com/events/7309948โ€ฆ

๐Ÿ’ฆ Everythingโ€™s fineโ€ฆ until it isnโ€™t. ๐Ÿ”ฅ
Generative AI runs on code and concrete, metals, heat, and vapor.
Letโ€™s talk about the real infrastructure behind your prompts. ๐Ÿ—โš™๏ธ๐Ÿ”ฅ๐Ÿ’จ
๐Ÿ—“ April 15, 5PM UTC
๐Ÿ”— linkedin.com/events/7309948โ€ฆ
Pruna AI (@prunaai) 's Twitter Profile Photo

๐—ฆ๐—ฎ๐˜† ๐—ต๐—ฒ๐˜† ๐˜๐—ผ ๐——๐—ฎ๐˜ƒ๐—ถ๐—ฑ ๐—•๐—ฒ๐—ฟ๐—ฒ๐—ป๐˜€๐˜๐—ฒ๐—ถ๐—ป (x.com/davidberenstei) โ€” ๐—ฃ๐—ฟ๐˜‚๐—ป๐—ฎโ€™๐˜€ ๐—ป๐—ฒ๐˜„ ๐——๐—ฒ๐˜ƒ๐—ฅ๐—ฒ๐—น ๐—ฎ๐—น๐—ฐ๐—ต๐—ฒ๐—บ๐—ถ๐˜€๐˜! ๐Ÿงช๐Ÿค– From Hugging Face & Argilla to homemade pasta ๐Ÿ, Davidโ€™s all about precision, flavor & flow! Now heโ€™s cooking with us at Pruna to make models

๐—ฆ๐—ฎ๐˜† ๐—ต๐—ฒ๐˜† ๐˜๐—ผ ๐——๐—ฎ๐˜ƒ๐—ถ๐—ฑ ๐—•๐—ฒ๐—ฟ๐—ฒ๐—ป๐˜€๐˜๐—ฒ๐—ถ๐—ป (x.com/davidberenstei) โ€” ๐—ฃ๐—ฟ๐˜‚๐—ป๐—ฎโ€™๐˜€ ๐—ป๐—ฒ๐˜„ ๐——๐—ฒ๐˜ƒ๐—ฅ๐—ฒ๐—น ๐—ฎ๐—น๐—ฐ๐—ต๐—ฒ๐—บ๐—ถ๐˜€๐˜! ๐Ÿงช๐Ÿค–
From Hugging Face & Argilla to homemade pasta ๐Ÿ, Davidโ€™s all about precision, flavor & flow!
Now heโ€™s cooking with us at Pruna to make models
Bertrand Charpentier (@bertrand_charp) 's Twitter Profile Photo

๐Ÿ’ฅFlux-juiced end point pushes the limits of image generation performance for both efficiency and quality!๐Ÿ’ฅ - Try Flux-Juiced on Replicate: lnkd.in/eBYCnhwg - Read the full benchmarking blog on Hugging Face : lnkd.in/epRWuidn

๐Ÿ’ฅFlux-juiced end point pushes the limits of image generation performance for both efficiency and quality!๐Ÿ’ฅ 

- Try Flux-Juiced on Replicate: lnkd.in/eBYCnhwg
- Read the full benchmarking blog on Hugging Face
: lnkd.in/epRWuidn
Bertrand Charpentier (@bertrand_charp) 's Twitter Profile Photo

๐—›๐—ผ๐˜„ ๐˜๐—ผ ๐—น๐—ฒ๐—ฎ๐—ฟ๐—ป ๐—ฎ๐—ฏ๐—ผ๐˜‚๐˜ ๐—ฒ๐—ณ๐—ณ๐—ถ๐—ฐ๐—ถ๐—ฒ๐—ป๐˜ ๐—”๐—œ? Happy to announce the Awesome AI Efficiency repo that gathers a ๐—ฐ๐˜‚๐—ฟ๐—ฎ๐˜๐—ฒ๐—ฑ ๐—น๐—ถ๐˜€๐˜ ๐—ผ๐—ณ ๐Ÿญ๐Ÿฌ๐Ÿฌ+ ๐—บ๐—ฎ๐˜๐—ฒ๐—ฟ๐—ถ๐—ฎ๐—น๐˜€ to understand the challenges and solutions in making AI faster, smaller, cheaper, greener. ๐Ÿš€ It is

Pruna AI (@prunaai) 's Twitter Profile Photo

๐Ÿ”ฅย Community: โ€Image editing is too damn slow!โ€ Donโ€™t worry, we accepted the challenge of making HiDream-e1 faster and it is now at 8,7s on an H100! :) โ˜•๏ธŽ Take a coffee, sit back and relax, as the model has been run over 20,000 times on Replicate already! ๐Ÿ‘‡ย Model URL in the

๐Ÿ”ฅย Community: โ€Image editing is too damn slow!โ€ Donโ€™t worry, we accepted the challenge of making HiDream-e1 faster and it is now at 8,7s on an H100! :)

โ˜•๏ธŽ Take a coffee, sit back and relax, as the model has been run over 20,000 times on <a href="/replicate/">Replicate</a> already!

๐Ÿ‘‡ย Model URL in the
Bertrand Charpentier (@bertrand_charp) 's Twitter Profile Photo

We made Flux-Kontext-dev from Black Forest Labs x5 faster in <4h and deployed in on Replicate. Hope that you will enjoy! We put details in our blog: pruna.ai/blog/flux-kontโ€ฆ :)

Pruna AI (@prunaai) 's Twitter Profile Photo

๐Ÿง‘โ€๐Ÿซย AI Efficiency Fundamentals - Week 1: Large Language Architectures Do you know the difference between Autoregressive, Diffusion, and State-Space LLMs? Even if you do, these slides are great for you. At Pruna, we want to educate about efficient AI, so our lead researcher and

๐Ÿง‘โ€๐Ÿซย AI Efficiency Fundamentals - Week 1: Large Language Architectures

Do you know the difference between Autoregressive, Diffusion, and State-Space LLMs? Even if you do, these slides are great for you.

At Pruna, we want to educate about efficient AI, so our lead researcher and
Bertrand Charpentier (@bertrand_charp) 's Twitter Profile Photo

How to make AI endpoints having less CO2 emissions? ๐ŸŒฑ One solution is to use endpoints with compressed models. This is particularly important when endpoints run at scale.

Pruna AI (@prunaai) 's Twitter Profile Photo

๐Ÿš€ ๐—ฃ๐—ฟ๐˜‚๐—ป๐—ฎ ๐˜… @๐—ด๐—ผ๐—ธ๐—ผ๐˜†๐—ฒ๐—ฏ ๐—ฃ๐—ฎ๐—ฟ๐˜๐—ป๐—ฒ๐—ฟ๐˜€๐—ต๐—ถ๐—ฝ ๐—จ๐—ฝ๐—ฑ๐—ฎ๐˜๐—ฒ! ๐Ÿ”ฅ Early adopters are reporting great results from our lightning-fast inference platform: Performance Breakthrough: โ€ข โšก๏ธ Much faster models โ€ข ๐Ÿ’ฐ Cost reduction โ€ข ๐ŸŽฏ Minimal quality degradation Letโ€™s talk

๐Ÿš€ ๐—ฃ๐—ฟ๐˜‚๐—ป๐—ฎ ๐˜… @๐—ด๐—ผ๐—ธ๐—ผ๐˜†๐—ฒ๐—ฏ ๐—ฃ๐—ฎ๐—ฟ๐˜๐—ป๐—ฒ๐—ฟ๐˜€๐—ต๐—ถ๐—ฝ ๐—จ๐—ฝ๐—ฑ๐—ฎ๐˜๐—ฒ!

๐Ÿ”ฅ Early adopters are reporting great results from our lightning-fast inference platform:

Performance Breakthrough:
โ€ข โšก๏ธ Much faster models
โ€ข ๐Ÿ’ฐ Cost reduction
โ€ข ๐ŸŽฏ Minimal quality degradation 

Letโ€™s talk
Bertrand Charpentier (@bertrand_charp) 's Twitter Profile Photo

From Wan video to Wan Image: We built the fastest endpoint for generating 2K images! - Accessible on Replicate : lnkd.in/eqsBR2Kx - Check details, examples, and benchmarks in our blog: lnkd.in/eXcAbqjM - Use Pruna AI to compress more AI models:

Pruna AI (@prunaai) 's Twitter Profile Photo

๐Ÿ”ฅ Deploy custom AI models with Pruna optimization speed + @lightningai LitServe serving engine! Lightning-Fast AI Deployments! What makes this awesome: โ€ข โšก๏ธ FastAPI-powered serving โ€ข ๐ŸŽฏ Built-in batching โ€ข โš™๏ธย Define and serve any model (vision, audio, text) โ€ข ๐Ÿš€ Easy

๐Ÿ”ฅ Deploy custom AI models with Pruna optimization speed + @lightningai LitServe serving engine! Lightning-Fast AI Deployments!

What makes this awesome:
โ€ข โšก๏ธ FastAPI-powered serving
โ€ข ๐ŸŽฏ Built-in batching
โ€ข โš™๏ธย Define and serve any model (vision, audio, text)
โ€ข ๐Ÿš€ Easy