CambridgeLTL (@cambridgeltl) 's Twitter Profile
CambridgeLTL

@cambridgeltl

Language Technology Lab (LTL) at the University of Cambridge. Computational Linguistics / Machine Learning / Deep Learning. Focus: Multilingual NLP and Bio NLP.

ID: 964208941977792512

linkhttp://ltl.mml.cam.ac.uk/ calendar_today15-02-2018 18:45:00

245 Tweet

2,2K Followers

85 Following

Yi Xu (@_yixu) 's Twitter Profile Photo

🚀Let’s Think Only with Images. No language and No verbal thought.🤔 Let’s think through a sequence of images💭, like how humans picture steps in their minds🎨. We propose Visual Planning, a novel reasoning paradigm that enables models to reason purely through images.

🚀Let’s Think Only with Images.

No language and No verbal thought.🤔 

Let’s think through a sequence of images💭, like how humans picture steps in their minds🎨. 

We propose Visual Planning, a novel reasoning paradigm that enables models to reason purely through images.
Benjamin Minixhofer (@bminixhofer) 's Twitter Profile Photo

We achieved the first instance of successful subword-to-byte distillation in our (just updated) paper. This enables creating byte-level models at a fraction of the cost of what was needed previously. As a proof-of-concept, we created byte-level Gemma2 and Llama3 models. đź§µ

We achieved the first instance of successful subword-to-byte distillation in our (just updated) paper.

This enables creating byte-level models at a fraction of the cost of what was needed previously.

As a proof-of-concept, we created byte-level Gemma2 and Llama3 models.

đź§µ
Caiqi Zhang (@caiqizh) 's Twitter Profile Photo

🔥 We teach LLMs to say how confident they are on-the-fly during long-form generation. 🤩No sampling. No slow post-hoc methods. Not limited to short-form QA! ‼️Just output confidence in a single decoding pass. ✅Better calibration! 🚀 20× faster runtime. arXiv:2505.23912 👇

🔥 We teach LLMs to say how confident they are on-the-fly during long-form generation.

🤩No sampling. No slow post-hoc methods. Not limited to short-form QA!

‼️Just output confidence in a single decoding pass.

âś…Better calibration!
🚀 20× faster runtime.

arXiv:2505.23912
👇
CambridgeLTL (@cambridgeltl) 's Twitter Profile Photo

Our LTL team had an amazing time at M2L school in Croatia 🇭🇷! 🏆 Hannah won Best Poster Award! → arxiv.org/html/2509.1481… 🎤 Ivan Vulić delivered an amazing lecture → drive.google.com/file/d/1feqnLe…

Our LTL team had an amazing time at <a href="/M2lSchool/">M2L school</a>  in Croatia 🇭🇷!
🏆 <a href="/h_sterz/">Hannah</a>  won Best Poster Award! → arxiv.org/html/2509.1481…
🎤 <a href="/licwu/">Ivan Vulić</a>  delivered an amazing lecture → drive.google.com/file/d/1feqnLe…
Tiancheng Hu (@tiancheng_hu) 's Twitter Profile Photo

Instruction tuning unlocks incredible skills in LLMs, but at a cost: they become dangerously overconfident. You face a choice: a well-calibrated base model or a capable but unreliable instruct model. What if you didn't have to choose? What if you could navigate the trade-off?

Sharan (@_maiush) 's Twitter Profile Photo

AI that is “forced to be good” v “genuinely good” Should we care about the difference? (yes!) We’re releasing the first open implementation of character training. We shape the persona of AI assistants in a more robust way than alternatives like prompting or activation steering.

AI that is “forced to be good” v “genuinely good”
Should we care about the difference? (yes!)

We’re releasing the first open implementation of character training. We shape the persona of AI assistants in a more robust way than alternatives like prompting or activation steering.