Mikel Artetxe (@artetxem) 's Twitter Profile
Mikel Artetxe

@artetxem

Co-founder @RekaAILabs and Honorary Researcher @IxaGroup (University of the Basque Country) | Past: Research Scientist @AIatMeta (FAIR)

ID: 892059194240532480

linkhttps://www.mikelartetxe.com calendar_today31-07-2017 16:27:40

582 Tweet

6,6K Followers

226 Following

Mikel Artetxe (@artetxem) 's Twitter Profile Photo

⚡️ Reka Flash ⚡️, our 21B multimodal assistant, is much better now 🔥🔥🔥 Try it for free at chat.reka.ai. No need to sign up anymore!

Reka (@rekaailabs) 's Twitter Profile Photo

Dani Yogatama, our CEO, had the opportunity to share about the work we do at Reka, our multimodal AI models, his perspectives of the future of AI, as well as how we use AMD Instinct MI300X and ROCm software at AMD's Advancing AI event.

Dani Yogatama, our CEO, had the opportunity to share about the work we do at Reka, our multimodal AI models, his perspectives of the future of AI, as well as how we use <a href="/AMD/">AMD</a>  Instinct MI300X and ROCm software at AMD's Advancing AI event.
Mikel Artetxe (@artetxem) 's Twitter Profile Photo

I wish there was a more direct question about keeping ARR vs. getting rid of it vs. making it optional, but it's still great that we have a chance to give our opinion. Please don't forget to vote!

Mikel Artetxe (@artetxem) 's Twitter Profile Photo

📢 Make multiple choice benchmarks more difficult by replacing a random choice with "None of the above"! Interesting to see how certain models are a lot more affected (cough cough Qwen 7B)... What does a stellar MMLU mean when such a trivial change causes a drop of 20 points? 🤔

Mikel Artetxe (@artetxem) 's Twitter Profile Photo

Looking for a challenging benchmark? BIG-Bench Extra Hard comprises 23 of them... and our linguistic reasoning benchmark, Linguini, is the hardest of them all, with all models scoring below 20!

Mikel Artetxe (@artetxem) 's Twitter Profile Photo

Open sourcing our 21B reasoning model! Trained from scratch, strong numbers, Apache 2.0... Go build with it! 🔥🔥🔥 And even more excited to announce Nexus, our new platform to create AI workers, available in private preview. More coming soon!

Reka (@rekaailabs) 's Twitter Profile Photo

Reka Flash 3 update: added another great model Gemma-3 27B (instruction tuned version) to our benchmark results since many people asked us how they compare on these datasets. Also Apache 2.0 vs Gemma license.

Reka Flash 3 update: added another great model Gemma-3 27B (instruction tuned version) to our benchmark results since many people asked us how they compare on these datasets.

Also Apache 2.0 vs Gemma license.
Xeophon (@thexeophon) 's Twitter Profile Photo

After multiple requests, here it finally is: Reka Flash 3 is an amazing model on my vibe eval! While its consistency is pretty bad, the pass@5 is crazy, on par with GFT in general and on par with Claude 3.5 in coding. And thats 21B w/ Apache 2.0 license!

After multiple requests, here it finally is: <a href="/RekaAILabs/">Reka</a> Flash 3 is an amazing model on my vibe eval!

While its consistency is pretty bad, the pass@5 is crazy, on par with GFT in general and on par with Claude 3.5 in coding. And thats 21B w/ Apache 2.0 license!
Mikel Artetxe (@artetxem) 's Twitter Profile Photo

Euskarazko ChatGPT ireki bat nahi? HiTZ zentroa (UPV/EHU)-n horretan ari gara eta azken bultzada emateko zure laguntza behar dugu! Oso erraza da parte hartzea: 1. 🌐Sartu ebaluatoia.hitz.eus helbidean 2. ❓Egin zure galdera 3. 🩷Aukeratu 2 erantzunen artean onena 4. 🎁 Irabazi sariak

Mikel Artetxe (@artetxem) 's Twitter Profile Photo

Another account of Reka Flash 3 doing strongly in the wild! 💪 "Reka Flash 3 is really capable, especially for its size. In our testing, it came close to the performance of 32B models, such as the R1-distill of Qwen and QwQ 32B"

Reka (@rekaailabs) 's Twitter Profile Photo

Vision capabilities are core to our mission to develop multimodal AI that understands both the digital and physical worlds. In our latest blogpost, we discuss our approach to process video inputs and our industry-leading results from different domains. reka.ai/news/reka-visu…

Mikel Artetxe (@artetxem) 's Twitter Profile Photo

Including English data when adapting LLMs to new languages doesn't affect perplexity, but can have a huge impact in downstream tasks 🤯 What causes this unusual uncorrelation? Why is English data so important? 🤔 Check out our new paper where we solve the mystery! 🧵👇

Including English data when adapting LLMs to new languages doesn't affect perplexity, but can have a huge impact in downstream tasks 🤯

What causes this unusual uncorrelation? Why is English data so important? 🤔 Check out our new paper where we solve the mystery! 🧵👇