Pieter Delobelle (@pieterdelobelle) 's Twitter Profile
Pieter Delobelle

@pieterdelobelle

Fairness in LLMs and Dutch NLP
Currently LLM engineer @aleph__alpha
Prev: @apple, PhD & postdoc @KU_Leuven

ID: 551272345

linkhttps://pieter.ai calendar_today11-04-2012 19:52:03

106 Tweet

376 Followers

480 Following

ludwig (@ludwigabap) 's Twitter Profile Photo

this guy's channel is so small with only a couple K views here and there if you're interested in GPU programming and still a beginner, he's worth a look (Simon Oz on yt)

this guy's channel is so small with only a couple K views here and there

if you're interested in GPU programming and still a beginner, he's worth a look 

(Simon Oz on yt)
Pieter Delobelle (@pieterdelobelle) 's Twitter Profile Photo

I also release some synthetic datasets I made with LLMQ by translating fineweb to Dutch and German. And with a permissive license (ODC-by). 🇩🇪 500k rows translated with Unbabel's Tower+ 72B: huggingface.co/datasets/pdelo… 🇳🇱 1.5M rows translated with Tower+ 9B huggingface.co/datasets/pdelo…

Pieter Delobelle (@pieterdelobelle) 's Twitter Profile Photo

Serving an LLM efficiently (=profitably) is highly non-trivial and involves a lot of different design choices. Mixture of experts, as used by Deepseek, complicates this a lot. I really learned to appreciate this from Piotr Mazurek while I was at Aleph Alpha, so check out this deep

Pablo Iyu Guerrero (@pabloiyu) 's Twitter Profile Photo

First high-performance inference for hierarchical byte models. Lukas Blübaum and I developed batched inference for tokenizer-free HAT (Hierarchical Autoregressive Transformers) models, developed by Aleph Alpha Research. In some settings, we outcompete the baseline Llama.🧵

First high-performance inference for hierarchical byte models.
<a href="/LukasBluebaum/">Lukas Blübaum</a> and I developed batched inference for tokenizer-free HAT (Hierarchical Autoregressive Transformers) models, developed by <a href="/Aleph__Alpha/">Aleph Alpha</a> Research. In some settings, we outcompete the baseline Llama.🧵