Yiyang Nan (@yiyangnan) 's Twitter Profile
Yiyang Nan

@yiyangnan

Research Scholar @Cohere_Labs | Prev: @BrownUniversity @UMich

ID: 1022433920023846912

linkhttp://nanyyyyyy.github.io calendar_today26-07-2018 10:50:17

26 Tweet

124 Followers

693 Following

Saurabh Dash (@theycallmemr_) 's Twitter Profile Photo

🚨Preprint Alert As promised, announcing the Aya Vision Technical Report – detailing the recipe to build SOTA multilingual multimodal models.

🚨Preprint Alert
As promised, announcing the Aya Vision Technical Report – detailing the recipe to build SOTA multilingual multimodal models.
Cohere Labs (@cohere_labs) 's Twitter Profile Photo

How do we build multimodal systems that work effectively across the globe? 🌍 Today we release the Aya Vision Technical Report, the detailed recipe behind Aya Vision models, unifying state-of-the-art multilingual capabilities in multimodal and text tasks across 23 languages!

Shan Chen (@shan23chen) 's Twitter Profile Photo

‼️ 1/n Ask your reasoning model to think in lower resource language does degrade models’ performance at the moment. My awesome Co-author already communicated the main points in the thread, I will just communicate some random things we learned in my 🧵

Yong Zheng-Xin (Yong) (@yong_zhengxin) 's Twitter Profile Photo

🧵 Multilingual safety training/eval is now standard practice, but a critical question remains: Is multilingual safety actually solved? Our new survey with Cohere Labs answers this and dives deep into: - Language gap in safety research - Future priority areas Thread 👇

🧵 Multilingual safety training/eval is now standard practice, but a critical question remains: Is multilingual safety actually solved?

Our new survey with <a href="/Cohere_Labs/">Cohere Labs</a> answers this and dives deep into:
- Language gap in safety research
- Future priority areas

Thread 👇
Brown CS (@browncsdept) 's Twitter Profile Photo

We're happy to announce that effective as of July 1, 2025, faculty members Stephen Bach and Srinath Sridhar have received named chairs. Steve is now the Eliot Horowitz Assistant Professor in CS and Srinath is the John E. Savage Assistant Professor in CS: cs.brown.edu/news/2025/06/0…

We're happy to announce that effective as of July 1, 2025, faculty members <a href="/stevebach/">Stephen Bach</a> and <a href="/drsrinathsridha/">Srinath Sridhar</a> have received named chairs. Steve is now the Eliot Horowitz Assistant Professor in CS and Srinath is the John E. Savage Assistant Professor in CS: cs.brown.edu/news/2025/06/0…
Stephen Bach (@stevebach) 's Twitter Profile Photo

Excited to release Trove, our lightweight, hackable, and easy-to-use toolkit for dense retrieval, which simplifies experiments. Trove makes it easy and efficient to explore all the combinations of your documents and queries for training, and streamlines multi-node evaluations.

Excited to release Trove, our lightweight, hackable, and easy-to-use toolkit for dense retrieval, which simplifies experiments.  Trove makes it easy and efficient to explore all the combinations of your documents and queries for training, and streamlines multi-node evaluations.
Stella Li (@stellalisy) 's Twitter Profile Photo

Excited to share more about Spurious Rewards! Also keep an eye out for some new experiments and arxiv coming soon 👀🔜

Cohere Labs (@cohere_labs) 's Twitter Profile Photo

How can we make language models more flexible to adapt to new languages after pretraining? 🌏 🧠 Our latest work investigates whether a tokenizer trained on more languages than the pretraining target can improve language plasticity without compromising pretraining performance.

How can we make language models more flexible to adapt to new languages after pretraining? 🌏

🧠 Our latest work investigates whether a tokenizer trained on more languages than the pretraining target can improve language plasticity without compromising pretraining performance.
Diana Abagyan (@dianaabagyan) 's Twitter Profile Photo

🚨New pretraining paper on multilingual tokenizers 🚨 Super excited to share my work with Cohere Labs: One Tokenizer To Rule Them All: Emergent Language Plasticity via Multilingual Tokenizers

🚨New pretraining paper on multilingual tokenizers 🚨

Super excited to share my work with <a href="/Cohere_Labs/">Cohere Labs</a>: One Tokenizer To Rule Them All: Emergent Language Plasticity via Multilingual Tokenizers
Sara Hooker (@sarahookr) 's Twitter Profile Photo

Huge congrats to Diana Abagyan on her first first author paper. Was a pleasure collaborating on this work — we ask what cheap interventions in pre-training can allow for more language plasticity downstream.

Sara Hooker (@sarahookr) 's Twitter Profile Photo

Thanks AK for the spotlight on our work I really believe strongly in this wider direction — of taking the pressure off everyday users to be master prompt engineers and inferring controllability directly from tasks.

Ahmet Üstün (@ahmetustun89) 's Twitter Profile Photo

Can we train models for better inference-time control instead of over-complex prompt engineering❓ Turns out the key is in the data — adding fine-grained markers boosts performance and enables flexible control at inference🎁 Huge congrats to Daniel D'souza  for this great work

Jack Clark (@jackclarksf) 's Twitter Profile Photo

Amir Efrati Stephanie Palazzolo Worth reading this research which showed it has already been turned into cheat-slop and that meta was one of the worst culprits for gaming it x.com/singhshiviii/s…

Yong Zheng-Xin (Yong) (@yong_zhengxin) 's Twitter Profile Photo

We see so many work this week about "emergent misalignment", but how is it fundamentally different from LLM jailbreaking research? I wrote a short blog post about it: yongzx.substack.com/p/emergent-mis…

We see so many work this week about "emergent misalignment", but how is it fundamentally different from LLM jailbreaking research?   

I wrote a short blog post about it:
yongzx.substack.com/p/emergent-mis…
Muhammad Khalifa (@mkhalifaaaa) 's Twitter Profile Photo

Last week, I gave a talk at Tsinghua University on scaling test-time compute with generative process verifiers/PRMs that verify reasoning with reasoning. jietang thank you for the invite! TL;DR: A super data-efficient recipe to train generative process verifiers that can scale

Last week, I gave a talk at Tsinghua University on scaling test-time compute with generative process verifiers/PRMs that verify reasoning with reasoning.

<a href="/jietang/">jietang</a> thank you for the invite!

TL;DR: A super data-efficient recipe to train generative process verifiers that can scale
Cohere Labs (@cohere_labs) 's Twitter Profile Photo

Proud to support South Korean research. 🇰🇷 Our new grant program is designed to encourage and support the Korean research ecosystem and social impact initiatives. Let’s innovate together.

Proud to support South Korean research. 🇰🇷

Our new grant program is designed to encourage and support the Korean research ecosystem and social impact initiatives.

Let’s innovate together.