Ercong Nie (@nielklug) 's Twitter Profile
Ercong Nie

@nielklug

PhD Student in Computational Linguistics & #NLProc @cislmu of @LMU_Muenchen, affiliated member @munichcenterML, Previously @sjtu1896

ID: 972522090225192961

linkhttps://www.cip.ifi.lmu.de/~nie/hpg.html calendar_today10-03-2018 17:18:29

382 Tweet

390 Followers

847 Following

Canyu Chen (@canyuchen3) 's Twitter Profile Photo

๐Ÿ”ฅNew #ICLR2025 paper: "๐‚๐š๐ง ๐Š๐ง๐จ๐ฐ๐ฅ๐ž๐๐ ๐ž ๐„๐๐ข๐ญ๐ข๐ง๐  ๐‘๐ž๐š๐ฅ๐ฅ๐ฒ ๐‚๐จ๐ซ๐ซ๐ž๐œ๐ญ ๐‡๐š๐ฅ๐ฅ๐ฎ๐œ๐ข๐ง๐š๐ญ๐ข๐จ๐ง๐ฌ?" Knowledge editing aims to fix false facts in LLMsโ€”but does it actually work? ๐Ÿค” We provide new insights!โœจ ๐Ÿ”—paper & code: llm-editing.github.io

๐Ÿ”ฅNew #ICLR2025 paper: "๐‚๐š๐ง ๐Š๐ง๐จ๐ฐ๐ฅ๐ž๐๐ ๐ž ๐„๐๐ข๐ญ๐ข๐ง๐  ๐‘๐ž๐š๐ฅ๐ฅ๐ฒ ๐‚๐จ๐ซ๐ซ๐ž๐œ๐ญ ๐‡๐š๐ฅ๐ฅ๐ฎ๐œ๐ข๐ง๐š๐ญ๐ข๐จ๐ง๐ฌ?" 

Knowledge editing aims to fix false facts in LLMsโ€”but does it actually work? ๐Ÿค” We provide new insights!โœจ

๐Ÿ”—paper & code: llm-editing.github.io
Yunzhi Yao (@yyztodd) 's Twitter Profile Photo

๐Ÿšจ New Blog Drop! ๐Ÿš€ "Reflection on Knowledge Editing: Charting the Next Steps" is live! ๐Ÿ’ก Ever wondered why knowledge editing in LLMs still feels more like a lab experiment than a real-world solution? In this post, we dive deep into where the research is thriving โ€” and where

Gabriele Sarti (@gsarti_) 's Twitter Profile Photo

BlackboxNLP is back! ๐Ÿ’ฅ Happy to be part of the organizing team for this year, and super excited for our new shared task using the excellent MIB Benchmark for circuit/causal variable localization in LMs, check it out! blackboxnlp.github.io/2025/task/

Ercong Nie (@nielklug) 's Twitter Profile Photo

Three papers accepted to ACL 2025๐Ÿฅฐ On cross-lingual knowledge editing, probing, and multilingual interpretability๐Ÿ” Sincere thanks to all the amazing coauthors and supervisors! โค๏ธ Looking forward to meeting old friends and knowing new people soon in Vienna! ๐Ÿ‡ฆ๐Ÿ‡น

Three papers accepted to ACL 2025๐Ÿฅฐ

On cross-lingual knowledge editing, probing, and multilingual interpretability๐Ÿ”

Sincere thanks to all the amazing coauthors and supervisors! โค๏ธ 

Looking forward to meeting old friends and knowing new people soon in Vienna! ๐Ÿ‡ฆ๐Ÿ‡น
Mingyang Wang โœˆ๏ธ ACL 2025 (@mingyang2666) 's Twitter Profile Photo

๐ŸŽ‰Excited to share our paper on cross-lingual inconsistency is accepted to #ACL2025 ๐Ÿ‡ฆ๐Ÿ‡น! We dissect why LLMs produce inconsistent outputs across languages using interpretability analysis, and propose a simple shortcut-based fix, evaluated on 17 languages. arxiv.org/abs/2504.04264

๐ŸŽ‰Excited to share our paper on cross-lingual inconsistency is accepted to #ACL2025 ๐Ÿ‡ฆ๐Ÿ‡น!

We dissect why LLMs produce inconsistent outputs across languages using interpretability analysis, and propose a simple shortcut-based fix, evaluated on 17 languages. arxiv.org/abs/2504.04264
Tiago Pimentel (@tpimentelms) 's Twitter Profile Photo

If you're finishing your camera-ready for ACL (#acl2025nlp) or ICML (#icml2025 ) and want to cite co-first authors more fairly, I just made a simple fix to do this! Just add $^*$ to the authors' names in your bibtex, and the citations should change :) github.com/tpimentelms/acโ€ฆ

If you're finishing your camera-ready for ACL (#acl2025nlp) or ICML (#icml2025 ) and want to cite co-first authors more fairly, I just made a simple fix to do this! Just add $^*$ to the authors' names in your bibtex, and the citations should change :)

github.com/tpimentelms/acโ€ฆ
Jirui Qi (@jirui_qi) 's Twitter Profile Photo

[1/]๐Ÿ’กNew Paper Large reasoning models (LRMs) are strong in English โ€” but how well do they reason in your language? Our latest work uncovers their limitation and a clear trade-off: Controlling Thinking Trace Language Comes at the Cost of Accuracy ๐Ÿ“„Link: arxiv.org/abs/2505.22888

[1/]๐Ÿ’กNew Paper
Large reasoning models (LRMs) are strong in English โ€” but how well do they reason in your language?

Our latest work uncovers their limitation and a clear trade-off:
Controlling Thinking Trace Language Comes at the Cost of Accuracy

๐Ÿ“„Link: arxiv.org/abs/2505.22888
Lucas Bandarkar (@lucasbandarkar) 's Twitter Profile Photo

The unreasonable effectiveness of model merging for cross-lingual transfer ! Our preprint evaluates a number of *modular* approaches to fine-tuning LLMs that "assign" model params to either task or language. Surprisingly, merging experts beats all ! ๐Ÿงต1/4 arxiv.org/abs/2505.18356

Amir H. Kargaran (@amir_nlp) 's Twitter Profile Photo

New paper: How does pretraining on programming languages + English shape LLMs' concept space? ๐Ÿ”Do LLMs use English or a programming language as a kind of pivot language? ๐Ÿง Are neurons language-specific or shared across programming languages and English? ๐Ÿ”—arxiv.org/abs/2506.01074

New paper: How does pretraining on programming languages + English shape LLMs' concept space?
๐Ÿ”Do LLMs use English or a programming language as a kind of pivot language?
๐Ÿง Are neurons language-specific or shared across programming languages and English?
๐Ÿ”—arxiv.org/abs/2506.01074
Mikel Artetxe (@artetxem) 's Twitter Profile Photo

Including English data when adapting LLMs to new languages doesn't affect perplexity, but can have a huge impact in downstream tasks ๐Ÿคฏ What causes this unusual uncorrelation? Why is English data so important? ๐Ÿค” Check out our new paper where we solve the mystery! ๐Ÿงต๐Ÿ‘‡

Including English data when adapting LLMs to new languages doesn't affect perplexity, but can have a huge impact in downstream tasks ๐Ÿคฏ

What causes this unusual uncorrelation? Why is English data so important? ๐Ÿค” Check out our new paper where we solve the mystery! ๐Ÿงต๐Ÿ‘‡
Weidi Xie (@weidixie) 's Twitter Profile Photo

For the soccer fans, don't miss this !!!!โšฝ๏ธโšฝ๏ธโšฝ๏ธโšฝ๏ธโšฝ๏ธ

Mor Geva (@megamor2) 's Twitter Profile Photo

โœจMLP layers have just become more interpretable than ever โœจ In a new paper: * We show a simple method for decomposing MLP activations into interpretable features * Our method uncovers hidden concept hierarchies, where sparse neuron combinations form increasingly abstract ideas

โœจMLP layers have just become more interpretable than ever โœจ
In a new paper:
* We show a simple method for decomposing MLP activations into interpretable features
* Our method uncovers hidden concept hierarchies, where sparse neuron combinations form increasingly abstract ideas
Arianna Bisazza (@ariannabisazza) 's Twitter Profile Photo

Happy to introduce TurBLiMP, the 1st benchmark of minimal pairs for free-order, morphologically rich Turkish! Pre-print: arxiv.org/abs/2506.13487 Fruit of an almost year-long project by amazing MS student Ezgi Basar, in collab w/ Francesca Padovani Jaap Jumelet

David Ifeoluwa Adelani ๐Ÿ‡ณ๐Ÿ‡ฌ (@davlanade) 's Twitter Profile Photo

Excited to announce the call for papers for the Multilingual Representation Learning workshop #EMNLP2025 sigtyp.github.io/ws2025-mrl.html with Duygu Ataman Catherine Arnett Jiayi Wang Fabian David Schmidt Tyler Chang Hila Gonen and amazing speakers: Alice Oh, Kelly Marchisio, & Pontus Stenetorp

Jixing Li (@jixingli) 's Twitter Profile Photo

Check out this super cool study on shared meaning representation across languages in brains and languages models ๐Ÿ‘‰๐Ÿป

Isabelle Augenstein (@iaugenstein) 's Twitter Profile Photo

๐ŸŽ“ Looking for PhD opportunities in #NLProc for a start in Spring 2026? ๐Ÿ—’๏ธ Add your expression of interest to join CopeNLU here by 20 July: forms.office.com/e/HZSmgR9nXB Selected candidates will be invited to submit a DARA fellowship application with me: daracademy.dk/fellowship/felโ€ฆ

ACL 2025 (@aclmeeting) 's Twitter Profile Photo

Exciting news! ๐ŸŽ‰ We're thrilled to announce an invited talk at ACL 2025 by Professor Luke Zettlemoyer (University of Washington and Meta)! ๐ŸŽค He'll be presenting on "Rethinking Pretraining: Data and Architecture." ๐Ÿง #ACL2025NLP #NLProc #InvitedTalk 2025.aclweb.org/program/keynotโ€ฆ