Tom Jobbins (@theblokeai) 's Twitter Profile
Tom Jobbins

@theblokeai

My Hugging Face repos: huggingface.co/TheBloke
Discord server: discord.gg/theblokeai
Patreon: patreon.com/TheBlokeAI

ID: 161524644

linkhttps://www.patreon.com/TheBlokeAI calendar_today01-07-2010 02:27:07

336 Tweet

15,15K Followers

229 Following

Tom Jobbins (@theblokeai) 's Twitter Profile Photo

Transformers 4.32.0 now supports GPTQ models natively! Over the last couple of days I have updated 296 of my GPTQ repos to provide automatic support for this. It's awesome you can now load a GPTQ model directly in Transformers with only two lines of code!

Tom Jobbins (@theblokeai) 's Twitter Profile Photo

Meta's CodeLlama is here! ai.meta.com/blog/code-llam… 7B, 7B-Instruct, 7B-Python, 13B, 13B-Instruct, 13B-Python, 34B, 34B-Instruct, 34B-Python First time we've seen the 34B model I've got a couple of fp16s up: huggingface.co/TheBloke/CodeL… huggingface.co/TheBloke/CodeL… More coming soon obvs

Tom Jobbins (@theblokeai) 's Twitter Profile Photo

Just released by Tav : Pygmalion 2, the sequel to one of the most popular models ever! And Mythalion, a new Gryphe merge! huggingface.co/TheBloke/Pygma… huggingface.co/TheBloke/Pygma… huggingface.co/TheBloke/Pygma… huggingface.co/TheBloke/Pygma… huggingface.co/TheBloke/Mytha… huggingface.co/TheBloke/Mytha…

Elinas (@officialelinas) 's Twitter Profile Photo

Chronos 70B v2 release! Thanks to Pygmalion for generously providing the compute and Tom Jobbins for quantizing the model. As usual, the model optimized for chat, roleplay, storywriting, and now includes vastly improved reasoning skills. huggingface.co/elinas/chronos…

Tom Jobbins (@theblokeai) 's Twitter Profile Photo

This is fantastic! Git clone was already dead for HF as far as I was concerned - I had my own hf_upload.py and hf_download.py scripts (wrapping HfAPI) for fast, efficient transfers. But huggingface_hub v0.17 makes those redundant! I will be using this now. Awesome stuff,🤗

lmarena.ai (formerly lmsys.org) (@lmarena_ai) 's Twitter Profile Photo

🔥Excited to introduce LMSYS-Chat-1M, a large-scale dataset of 1M real-world conversations with 25 cutting-edge LLMs! This dataset, collected from chat.lmsys.org, offers insights into user interactions with LLMs and intriguing use cases. Link: huggingface.co/datasets/lmsys…

younes (@younesbelkada) 's Twitter Profile Photo

New feature alert in the Hugging Face ecosystem! Flash Attention 2 natively supported in huggingface transformers, supports training PEFT, and quantization (GPTQ, QLoRA, LLM.int8) First pip install flash attention and pass use_flash_attention_2=True when loading the model!

New feature alert in the <a href="/huggingface/">Hugging Face</a> ecosystem!

Flash Attention 2 natively supported in huggingface transformers, supports training PEFT, and quantization (GPTQ, QLoRA, LLM.int8)

First pip install flash attention and pass use_flash_attention_2=True when loading the model!
Tom Jobbins (@theblokeai) 's Twitter Profile Photo

Thanks again to Latitude.sh for the loan of a beast 8xH100 server this week. I uploaded over 550 new repos, maybe my busiest week yet! Quanting is really resource intensive. Needs not only fast GPUs, but many CPUs, lots of disk, and 🚀 network. A server that ✅ all is v. rare!

Julien Chaumond (@julien_c) 's Twitter Profile Photo

oh hello Tom Jobbins I want to bookmark your 'Recent models' Collection on Hugging Face 🔥 Well... you can now upvote Collections! and browse upvoted collections on your profile ❤️

oh hello <a href="/TheBlokeAI/">Tom Jobbins</a> 

I want to bookmark your 'Recent models' Collection on <a href="/huggingface/">Hugging Face</a> 🔥

Well... you can now upvote Collections!

and browse upvoted collections on your profile ❤️
Victor M (@victormustar) 's Twitter Profile Photo

🤔 Are you interested in a "Follow" feature on the Hugging Face Hub? ➡️ This will allow you to see new models/records/spaces from users you follow.

🤔 Are you interested in a "Follow" feature on the Hugging Face Hub?
➡️ This will allow you to see new models/records/spaces from users you follow.
Tom Jobbins (@theblokeai) 's Twitter Profile Photo

It's been awesome to see Transformers getting support for more and more quantisation methods. And I've loved collaborating with younes and Hugging Face again! All my AWQ uploads now support Transformers. READMEs will update soon to show a Transformers Python example.

younes (@younesbelkada) 's Twitter Profile Photo

Blazing fast text generation using AWQ and fused modules! 🚀 Up to 3x speedup compared to native fp16 that you can use right now on any models supported by Tom Jobbins Simply pass an `AwqConfig` with `do_fuse=True` to `from_pretrained` method! huggingface.co/docs/transform…

Aleksa Gordić (水平问题) (@gordic_aleksa) 's Twitter Profile Photo

Tom Jobbins joined me to share his work in the open-source AI space - don't miss it! happening right now server link: discord.gg/peBrCpheKE (see the general channel or events channel for google meet link)

<a href="/TheBlokeAI/">Tom Jobbins</a> joined me to share his work in the open-source AI space - don't miss it! happening right now

server link: discord.gg/peBrCpheKE

(see the general channel or events channel for google meet link)
Tom Jobbins (@theblokeai) 's Twitter Profile Photo

Transformers now supports Mixtral GPTQs and I've updated my READMEs accordingly. It was awesome working with Marc Sun and younes of Hugging Face on this! Credit to LaaZa for coding the AutoGPTQ quant and inference implementation which enabled me to get GPTQs out fast!

emozilla (@theemozilla) 's Twitter Profile Photo

FYI to anyone using Mistral AI's Mixtral for long context tasks -- you can get even better performance by disabling sliding window attention (setting it to your max context length) config.sliding_window = 32768

FYI to anyone using <a href="/MistralAI/">Mistral AI</a>'s Mixtral for long context tasks -- you can get even better performance by disabling sliding window attention (setting it to your max context length)

config.sliding_window = 32768