Ivan Vulić (@licwu) 's Twitter Profile
Ivan Vulić

@licwu

PRA@Cambridge; Interested in (way) too many things for his well-being, but mostly (and rarely) (re)tweets about NLP, ML, IR, language(s); (likes parentheses)

ID: 930077097028145154

linkhttps://sites.google.com/site/ivanvulic/ calendar_today13-11-2017 14:17:14

202 Tweet

2,2K Followers

330 Following

Edoardo Ponti (@pontiedoardo) 's Twitter Profile Photo

We scaled sparse fine-tuning (SFT) to LLMs (such as Llama 2) by making it both parameter- and memory-efficient! (q)SFT instruction tuning performance is often better than (q)LoRA with comparable speed and memory load. Paper: arxiv.org/abs/2401.16405 Code:

Ivan Vulić (@licwu) 's Twitter Profile Photo

Think globally, act locally? Well, we were thought-experimenting whether LLMs would understand people from different places around our hometowns better than we ever might... And then we have eventually decided to make an actual (non-thought) experiment out of these thoughts! 👇👇

Edoardo Ponti (@pontiedoardo) 's Twitter Profile Photo

I am still looking for PhD students starting in September 2024! The deadline to apply for the CDT in NLP is the 11th of March. If you wish to do research in modular and efficient LLMs, here are some highlights of my lab's research from the past year ⬇️🧵

Sebastian Ruder (@seb_ruder) 's Twitter Profile Photo

🚨 A belated update: Our survey on "Modular Deep Learning" has been published in TMLR. Check out the updated version: openreview.net/forum?id=z9EkX…

Ivan Vulić (@licwu) 's Twitter Profile Photo

If we align LLMs through preferences, perhaps we should also evaluate them the same way (and respect transitivity)? The answer is: yes, we should. The trick, however, is how to make evaluation tractable. If you are into the whole "LLM-as-Judges" line of work, check this paper!

Neil Houlsby (@neilhoulsby) 's Twitter Profile Photo

Adapters are just a great way to share/benefit from new capabilities without handing around the kitchen sink. Congrats to the AdapterHub folks for adding support for quantized training (Q-LoRA and friends).

Benjamin Minixhofer (@bminixhofer) 's Twitter Profile Photo

Introducing Zero-Shot Tokenizer Transfer (ZeTT) ⚡ ZeTT frees language models from their tokenizer, allowing you to use any model with any tokenizer, with little or no extra training. Super excited to (finally!) share the first project of my PhD🧵

Introducing Zero-Shot Tokenizer Transfer (ZeTT) ⚡

ZeTT frees language models from their tokenizer, allowing you to use any model with any tokenizer, with little or no extra training.

Super excited to (finally!) share the first project of my PhD🧵
Chengzu Li (@li_chengzu) 's Twitter Profile Photo

Excited to introduce TopViewRS: VLMs as Top-View Spatial Reasoners🤖 TopViewRS assess VLMs’ spatial reasoning in top-view scenarios🏠just like how you read maps🗺️ Spoiler🫢GPT4V and Gemini are neck-and-neck, each excelling in different setups but neither even close to us humans

Excited to introduce TopViewRS: VLMs as Top-View Spatial Reasoners🤖

TopViewRS assess VLMs’ spatial reasoning in top-view scenarios🏠just like how you read maps🗺️

Spoiler🫢GPT4V and Gemini are neck-and-neck, each excelling in different setups but neither even close to us humans
Han Zhou (@hanzhou032) 's Twitter Profile Photo

Which output is better? [A] or [B]? LLM🤖: B❌ [B] or [A]? LLM🤖: A✅ Thrilled to share our preprint in addressing preference biases in LLM judgments!🧑‍⚖️We introduce ZEPO, a 0-shot prompt optimizer that enhances your LLM evaluators via fairness⚖️ 📰Paper: arxiv.org/abs/2406.11370

Which output is better?
[A] or [B]? LLM🤖: B❌
[B] or [A]? LLM🤖: A✅

Thrilled to share our preprint in addressing preference biases in LLM judgments!🧑‍⚖️We introduce ZEPO, a 0-shot prompt optimizer that enhances your LLM evaluators via fairness⚖️

📰Paper: arxiv.org/abs/2406.11370
Ivan Vulić (@licwu) 's Twitter Profile Photo

As someone who spent years working in multilingual NLP, I am so happy that we're finally seeing (L)LMs and (N)MT systems working in tandem towards the shared cause. The idea in this work is so simple & sweet, and yet it moves! 🌍🌏🌎

Markus Frohmann (@frohmannm) 's Twitter Profile Photo

Introducing 🪓Segment any Text! 🪓 A new state-of-the-art sentence segmentation tool! Compared to existing tools (and strong LLMs!), our models are far more: 1. efficient ⚡ 2. performant 🔝 3. robust 🚀 4. adaptable 🎯 5. multilingual 🗺

Introducing 🪓Segment any Text! 🪓

A new state-of-the-art sentence segmentation tool!
Compared to existing tools (and strong LLMs!), our models are far more:
1. efficient ⚡
2. performant 🔝
3. robust 🚀
4. adaptable 🎯
5. multilingual 🗺
Hannah (@h_sterz) 's Twitter Profile Photo

Do you DARE? Introducing a multiple-choice VQA benchmark ✨DARE✨ with: - 4 main robustness evaluation ⛓️ - 5 diverse categories 🧩 - Extensive analysis of 4 widely used VLMS 🤖

River Yijiang Dong (@river_dong121) 's Twitter Profile Photo

Thrilled to share our updated paper: "UNDIAL: Self-Distillation with Adjusted Logits for Robust Unlearning in Large Language Models" We propose a new robust LLM unlearning method via Self-Distillation on Adjusted Logits (UNDIAL). 📄 Paper: arxiv.org/pdf/2402.10052

Thrilled to share our updated paper: "UNDIAL: Self-Distillation with Adjusted Logits for Robust Unlearning in Large Language Models"
We propose a new robust LLM unlearning method via Self-Distillation on Adjusted Logits (UNDIAL).
📄 Paper: arxiv.org/pdf/2402.10052
Fabian David Schmidt (@fdschmidt) 's Twitter Profile Photo

📣Happy to (pre-)release my Fleurs-SLU benchmark to evaluate massively multilingual spoken language understanding on SIB & Belebele. Work done at Mila - Institut québécois d'IA with David Ifeoluwa Adelani 🇳🇬 Goran Glavaš Ivan Vulić Datasets: huggingface.co/datasets/WueNL… huggingface.co/datasets/WueNL… Details to follow👇

Benjamin Minixhofer (@bminixhofer) 's Twitter Profile Photo

We created Approximate Likelihood Matching, a principled (and very effective) method for *cross-tokenizer distillation*! With ALM, you can create ensembles of models from different families, convert existing subword-level models to byte-level and a bunch more🧵

We created Approximate Likelihood Matching, a principled (and very effective) method for *cross-tokenizer distillation*!

With ALM, you can create ensembles of models from different families, convert existing subword-level models to byte-level and a bunch more🧵
Yi Xu (@_yixu) 's Twitter Profile Photo

🚀Let’s Think Only with Images. No language and No verbal thought.🤔 Let’s think through a sequence of images💭, like how humans picture steps in their minds🎨. We propose Visual Planning, a novel reasoning paradigm that enables models to reason purely through images.

🚀Let’s Think Only with Images.

No language and No verbal thought.🤔 

Let’s think through a sequence of images💭, like how humans picture steps in their minds🎨. 

We propose Visual Planning, a novel reasoning paradigm that enables models to reason purely through images.
Benjamin Minixhofer (@bminixhofer) 's Twitter Profile Photo

We achieved the first instance of successful subword-to-byte distillation in our (just updated) paper. This enables creating byte-level models at a fraction of the cost of what was needed previously. As a proof-of-concept, we created byte-level Gemma2 and Llama3 models. 🧵

We achieved the first instance of successful subword-to-byte distillation in our (just updated) paper.

This enables creating byte-level models at a fraction of the cost of what was needed previously.

As a proof-of-concept, we created byte-level Gemma2 and Llama3 models.

🧵
Han Zhou (@hanzhou032) 's Twitter Profile Photo

Automating Multi-Agent Design: 🧩Multi-agent systems aren’t just about throwing more LLM agents together. 🛠️They require mastering the subtle art of prompting and agent orchestration. Introducing MASS🚀- Our new agent optimization framework for better prompts and topologies!

Automating Multi-Agent Design:

🧩Multi-agent systems aren’t just about throwing more LLM agents together.

🛠️They require mastering the subtle art of prompting and agent orchestration.

Introducing MASS🚀- Our new agent optimization framework for better prompts and topologies!
Lucas Caccia (@lucaspcaccia) 's Twitter Profile Photo

RAG and in-context learning are the go-to approaches for integrating new knowledge into LLMs, making inference very inefficient We propose instead 𝗞𝗻𝗼𝘄𝗹𝗲𝗱𝗴𝗲 𝗠𝗼𝗱𝘂𝗹𝗲𝘀 : lightweight LoRA modules trained offline that can match RAG performance without the drawbacks