Workshop on Large Language Model Memorization (@l2m2_workshop) 's Twitter Profile
Workshop on Large Language Model Memorization

@l2m2_workshop

The First Workshop on Large Language Model Memorization.

ID: 1835618506651373568

linkhttps://sites.google.com/view/memorization-workshop/home calendar_today16-09-2024 09:55:43

18 Tweet

100 Followers

12 Following

Workshop on Large Language Model Memorization (@l2m2_workshop) 's Twitter Profile Photo

🎉 Happy to announce that the L2M2 workshop has been accepted at ACL 2025! #NLProc #ACL2025 More details will follow soon. Stay tuned and spread the word! 📣

Niloofar (on faculty job market!) (@niloofar_mire) 's Twitter Profile Photo

Adding or removing PII in LLM training can *unlock previously unextractable* info. Even if “John.Mccarthy” never reappears, enough Johns & Mccarthys during post-training can make it extractable later! New paper on PII memorization & n-gram overlaps: arxiv.org/abs/2502.15680

Adding or removing PII in LLM training can *unlock previously unextractable* info. 

Even if “John.Mccarthy” never reappears, enough Johns & Mccarthys during post-training can make it extractable later! 

New paper on PII memorization & n-gram overlaps:
arxiv.org/abs/2502.15680
Ashwinee Panda (@pandaashwinee) 's Twitter Profile Photo

we show for the first time ever how to privacy audit LLM training. we give new SOTA methods that show how much models can memorize. by using our methods, you can know beforehand whether your model is going to memorize its training data, and how much, and when, and why! (1/n 🧵)

we show for the first time ever how to privacy audit LLM training. we give new SOTA methods that show how much models can memorize. by using our methods, you can know beforehand whether your model is going to memorize its training data, and how much, and when, and why! (1/n 🧵)
Abhilasha Ravichander (@lasha_nlp) 's Twitter Profile Photo

Want to know what training data has been memorized by models like GPT-4? We propose information-guided probes, a method to uncover memorization evidence in *completely black-box* models, without requiring access to 🙅‍♀️ Model weights 🙅‍♀️ Training data 🙅‍♀️ Token probabilities 🧵1/5

Want to know what training data has been memorized by models like GPT-4?

We propose information-guided probes, a method to uncover memorization evidence in *completely black-box* models,

without requiring access to
🙅‍♀️ Model weights
🙅‍♀️ Training data
🙅‍♀️ Token probabilities 🧵1/5
Workshop on Large Language Model Memorization (@l2m2_workshop) 's Twitter Profile Photo

Hi all, reminder that our direct submission deadline is April 15th! We are co-located at ACL'25 and you can submit archival or non-archival. You can also submit work published elsewhere (non-archival) Hope to see your submission! sites.google.com/view/memorizat…

Jiacheng Liu (@liujc1998) 's Twitter Profile Photo

As infini-gram surpasses 500 million API calls, today we're announcing two exciting updates: 1. Infini-gram is now open-source under Apache 2.0! 2. We indexed the training data of OLMo 2 models. Now you can search in the training data of these strong, fully-open LLMs. 🧵 (1/4)

Ai2 (@allen_ai) 's Twitter Profile Photo

For years it’s been an open question — how much is a language model learning and synthesizing information, and how much is it just memorizing and reciting? Introducing OLMoTrace, a new feature in the Ai2 Playground that begins to shed some light. 🔦

Tom McCoy (@rtommccoy) 's Twitter Profile Photo

Do language models just copy text they've seen before, or do they have generalizable abilities? ⬇️This new tool from Ai2 will be very useful for such questions! And allow me to plug our paper on this topic: We find that LLMs are mostly not copying! direct.mit.edu/tacl/article/d… 1/2

Do language models just copy text they've seen before, or do they have generalizable abilities?

⬇️This new tool from Ai2 will be very useful for such questions!

And allow me to plug our paper on this topic: We find that LLMs are mostly not copying!
direct.mit.edu/tacl/article/d…

1/2
Workshop on Large Language Model Memorization (@l2m2_workshop) 's Twitter Profile Photo

L2M2 is happening this Friday in Vienna at ACL 2025 #ACL2025NLP! We look forward to the gathering of memorization researchers in the NLP community. Invited talks include: Yanai Elazar Niloofar Reza Shokri and see our website for the full program. sites.google.com/view/memorizat…

Niloofar (on faculty job market!) (@niloofar_mire) 's Twitter Profile Photo

I'm psyched for my 2 *different* talks on Friday ACL 2025: 1.LLM Security (11:00): What does it mean for an AI agent to preserve privacy? 2.Workshop on Large Language Model Memorization (16:00): Emergent Misalignment thru the Lens of Non-verbatim Memorization (& phonetic to visual attacks!) Join us!

I'm psyched for my 2 *different* talks on Friday <a href="/aclmeeting/">ACL 2025</a>: 

1.<a href="/llm_sec/">LLM Security</a> (11:00): What does it mean for an AI agent to preserve privacy? 
  
2.<a href="/l2m2_workshop/">Workshop on Large Language Model Memorization</a> (16:00): Emergent Misalignment thru the Lens of Non-verbatim Memorization (&amp; phonetic to visual attacks!) 

Join us!
Workshop on Large Language Model Memorization (@l2m2_workshop) 's Twitter Profile Photo

L2M2 will be tomorrow at VIC, room 1.31-32! We hope you will join us for a day of invited talks, orals, and posters on LLM memorization. The full schedule and accepted papers are now on our website: sites.google.com/view/memorizat…