Giovanni Monea (@giomonea) 's Twitter Profile
Giovanni Monea

@giomonea

🤖 NLP PhD student @cornell @cornell_tech | Previously @Apple, @amazon, @EPFL_en, @polimi

ID: 1681411154524946446

linkhttp://giovannimonea.com calendar_today18-07-2023 21:10:41

41 Tweet

196 Followers

212 Following

Mustafa Omer Gul (@momergul_) 's Twitter Profile Photo

New paper! Models that learn from feedback train on their own outputs, so you see performance 📈 but language diversity 📉. We show that if you couple comprehension and generation you learn faster 🏎️ AND get richer language! arxiv.org/abs/2408.15992 Demo and video ⬇ + in EMNLP!

Julian Minder (@jkminder) 's Twitter Profile Photo

Can we understand and control how language models balance context and prior knowledge? Our latest paper shows it’s all about a 1D knob! 🎛️ arxiv.org/abs/2411.07404 Co-led with Kevin Du, as well as Niklas Stoehr, Giovanni Monea, Chris Wendler, Bob West & Ryan Cotterell.

Mustafa Omer Gul (@momergul_) 's Twitter Profile Photo

This still feels very surreal! I would like to thank EMNLP 2025 for this great honor, Yoav Artzi and my labmates for all of their support, and the many crowdworkers who played with and provided the feedback for our models!

Kianté Brantley (@xkianteb) 's Twitter Profile Photo

I am recruiting PhD students to join my lab at Harvard in Fall 2025! (deadline Dec 15) If you are interested in solving problems at the intersection of reinforcement learning, imitation learning, and NLP, pls consider applying (bit.ly/4fnficx)! Harvard SEAS Kempner Institute at Harvard University

Oreva Ahia (@orevaahia) 's Twitter Profile Photo

I am excited to be presenting MAGNET 🧲at NeurIPS 2024 next week. Subword tokenizers have been shown to overly segment text in non-Latin script languages. Our work presents an approach to train tokenizer-free multilingual LMs via efficient byte-level modeling. 1/n

I am excited to be presenting MAGNET 🧲at NeurIPS 2024 next week. Subword tokenizers have been shown to overly segment text in non-Latin script languages. 
Our work presents an approach to train tokenizer-free multilingual LMs via efficient byte-level modeling. 
1/n
Chris Wendler (@wendlerch) 's Twitter Profile Photo

SAEs pick up on abstract grammatical concepts that LLMs share across a diverse set of languages - even languages in which these grammatical concepts manifest in wildly different forms 🐍 🐈‍⬛ 🐮 see thread below 👇

Yoav Artzi (@yoavartzi) 's Twitter Profile Photo

I am looking for a postdoc. A serious-looking call coming soon, but this is to get it going. Topics include (but not limited to): LLMs (🫢!), multimodal LLMs, interaction+learning, RL, intersection with cogsci, ... see our work to get an idea: yoavartzi.com/pubs Plz RT 🙏

Yoav Artzi (@yoavartzi) 's Twitter Profile Photo

We now have a form for postdoc applications: forms.gle/tiydAChgV1wLcQ… I am looking at candidates on a rolling basis, so while there's no deadline, there's an advantage of throwing your name in the ring earlier than later

Yoav Artzi (@yoavartzi) 's Twitter Profile Photo

We recently pushed an update to this paper. Usually, updates don't justify a post, but this one is exceptionally contentful -> 🧵 tldr: all the findings are stronger, and the behaviors are super cool! arxiv.org/abs/2410.05362

We recently pushed an update to this paper. Usually, updates don't justify a post, but this one is exceptionally contentful -> 🧵

tldr: all the findings are stronger, and the behaviors are super cool!

arxiv.org/abs/2410.05362
Conference on Language Modeling (@colm_conf) 's Twitter Profile Photo

A bit of a mess around the conflict of COLM with the ARR (and to lesser degree ICML) reviews release. We feel this is creating a lot of pressure and uncertainty. So, we are pushing our deadlines: Abstracts due March 22 AoE (+48hr) Full papers due March 28 AoE (+24hr) Plz RT 🙏

A bit of a mess around the conflict of COLM with the ARR (and to lesser degree ICML) reviews release. We feel this is creating a lot of pressure and uncertainty. So, we are pushing our deadlines:

Abstracts due March 22 AoE (+48hr)
Full papers due March 28 AoE (+24hr)

Plz RT 🙏
Anthropic (@anthropicai) 's Twitter Profile Photo

How does Claude understand different languages? We find shared circuitry underlying the same concepts in multiple languages, implying that Claude "thinks" using universal concepts even before converting those thoughts into language.

How does Claude understand different languages? We find shared circuitry underlying the same concepts in multiple languages, implying that Claude "thinks" using universal concepts even before converting those thoughts into language.
Veniamin Veselovsky (@vminvsky) 's Twitter Profile Photo

New paper: Language models have “universal” concept representation – but can they capture cultural nuance? 🌏 If someone from Japan asks an LLM what color a pumpkin is, will it correctly say green (as they are in Japan)? Or does cultural nuance require more than just language?

New paper: Language models have “universal” concept representation – but can they capture cultural nuance? 🌏

If someone from Japan asks an LLM what color a pumpkin is, will it correctly say green (as they are in Japan)?

Or does cultural nuance require more than just language?
Rishi Jha (@rishi_d_jha) 's Twitter Profile Photo

I’m stoked to share our new paper: “Harnessing the Universal Geometry of Embeddings” with jack morris, Collin Zhang, and Vitaly Shmatikov. We present the first method to translate text embeddings across different spaces without any paired data or encoders. Here's why we're excited: 🧵👇🏾

I’m stoked to share our new paper: “Harnessing the Universal Geometry of Embeddings” with <a href="/jxmnop/">jack morris</a>, Collin Zhang, and <a href="/shmatikov/">Vitaly Shmatikov</a>.

We present the first method to translate text embeddings across different spaces without any paired data or encoders.

Here's why we're excited: 🧵👇🏾