Teknium (e/λ) (@teknium1) 's Twitter Profile
Teknium (e/λ)

@teknium1

Cofounder and Head of Post Training @NousResearch, prev @StabilityAI
Github: github.com/teknium1
HuggingFace: huggingface.co/teknium

ID: 1365020011123773442

linkhttp://github.com/sponsors/teknium1 calendar_today25-02-2021 19:25:11

38,38K Tweet

45,45K Followers

3,3K Following

Teknium (e/λ) (@teknium1) 's Twitter Profile Photo

With classifiers, whats the rule of thumb on label distribution? Do you want i.e. on a binary classifier to have 2 equally diverse distributions of 0's and 1's (an overall 50/50 split) or do you want it closer to the real world expected distribution of the binary labels (maybe

Teknium (e/λ) (@teknium1) 's Twitter Profile Photo

If an llm can use web search/browse during evals, and the evals are on the internet… cant they just search up the test set and find the answer? Lol

Teknium (e/λ) (@teknium1) 's Twitter Profile Photo

Imo cursor should just make a plan thats 15$ a month to get access to all the features and just be pay api pricing for everything it seems silly to prepay your api prices in advanced when api pricing is specifically because you pay per use and that varies

Teknium (e/λ) (@teknium1) 's Twitter Profile Photo

If you were to annotate and label a post training/instruct/chat dataset, what annotations do you think would be valuable for filtering, ablating, and testing subsets made out of? For example, a label of “refusal” on every sequence would be useful to filter with to reduce

Teknium (e/λ) (@teknium1) 's Twitter Profile Photo

I see all the stuff Janus and crew do with opus and gpt4 and I still aint gotten psychosis I must be doing something wrong..

emozilla (@theemozilla) 's Twitter Profile Photo

MoE money, MoE problems: it's straight up bonkers that there is not a single finetune of llama 4. zero. zilch. nada. everything on the hub is a reupload. trust me, I've spent the past several weeks trying with torchtune, torchtitan, hf -- anything. it literally just doesn't

MoE money, MoE problems:

it's straight up bonkers that there is not a single finetune of llama 4. 

zero. zilch. nada. everything on the hub is a reupload.

trust me, I've spent the past several weeks trying with torchtune, torchtitan, hf -- anything. it literally just doesn't
Teknium (e/λ) (@teknium1) 's Twitter Profile Photo

How long of generations can Gemini 2.5 make - I need it to reorganize 3200 lines of text and I dont think it can write this much lol

Teknium (e/λ) (@teknium1) 's Twitter Profile Photo

Answer: about 500 lines, before Gemini Chat just deletes its whole response and says you stopped it. Theoretically much longer outputs than anything but OAI's DeepResearch, but the interface ruins it. Maybe the API is better.