Fei Wang (@fwang_nlp) 's Twitter Profile
Fei Wang

@fwang_nlp

PhD candidate @USC.
Student Researcher @Google.
ML PhD Fellow @Amazon.
Trustworthy (Multimodal) LLMs.

ID: 1131226779413241856

linkhttps://feiwang96.github.io/ calendar_today22-05-2019 15:54:16

268 Tweet

1,1K Followers

2,2K Following

Bowen Jin (@bowenjin13) 's Twitter Profile Photo

LLM Alignment as Retriever Optimization: An Information Retrieval Perspective arxiv.org/abs/2502.03699 We introduce a comprehensive framework that connects LLM alignment techniques with the established IR principles, providing a new perspective on LLM alignment.

LLM Alignment as Retriever Optimization: An Information Retrieval Perspective
arxiv.org/abs/2502.03699
We introduce a comprehensive framework that connects LLM alignment techniques with the established IR principles, providing a new perspective on LLM alignment.
Kai-Wei Chang (@kaiwei_chang) 's Twitter Profile Photo

I'm honored to be named to “AI's 10 to Watch” by IEEE Intelligent Systems for our research on Trustworthy NLP and Vision-Language Models. This biennial award recognizes 10 researchers who have made impactful scientific contributions to AI within 10 years of earning their PhD.

Sheng Zhang (@sheng_zh) 's Twitter Profile Photo

I know internship hunting has been especially tough this year -- I hear you! 📢Great news: Our team at Microsoft Research (Microsoft Research) has opened up a few additional internship slots for this summer! We're seeking talented PhD candidates with experience in 🖼️Multimodality,

Sheng Zhang (@sheng_zh) 's Twitter Profile Photo

🚀 Excited to share MetaScale, our latest work advancing LLM reasoning capabilities! MetaScale empowers GPT-4o to match or even surpass frontier reasoning models like o1, Claude-3.5-Sonnet, and o1-mini on the challenging Arena-Hard benchmark (lmarena.ai). Additionally, MetaScale

🚀 Excited to share MetaScale, our latest work advancing LLM reasoning capabilities! MetaScale empowers GPT-4o to match or even surpass frontier reasoning models like o1, Claude-3.5-Sonnet, and o1-mini on the challenging Arena-Hard benchmark (<a href="/lmarena_ai/">lmarena.ai</a>). Additionally, MetaScale
Tu Vu (@tuvllms) 's Twitter Profile Photo

🚨 New paper 🚨 Excited to share my first paper w/ my PhD students!! We find that advanced LLM capabilities conferred by instruction or alignment tuning (e.g., SFT, RLHF, DPO, GRPO) can be encoded into model diff vectors (à la task vectors) and transferred across model

🚨 New paper 🚨

Excited to share my first paper w/ my PhD students!!

We find that advanced LLM capabilities conferred by instruction or alignment tuning (e.g., SFT, RLHF, DPO, GRPO) can be encoded into model diff vectors (à la task vectors) and transferred across model
Tianyi Lorena Yan (@lorenayannnnn) 's Twitter Profile Photo

When answering queries with multiple answers (e.g., listing cities of a country), how do LMs simultaneously recall knowledge and avoid repeating themselves? 🚀 Excited to share our latest work with Robin Jia! We uncover a promote-then-suppress mechanism: LMs first recall all

When answering queries with multiple answers (e.g., listing cities of a country), how do LMs simultaneously recall knowledge and avoid repeating themselves?

🚀 Excited to share our latest work with <a href="/robinomial/">Robin Jia</a>! We uncover a promote-then-suppress mechanism: LMs first recall all
Fei Liu @ #ICLR2025 (@feiliu_nlp) 's Twitter Profile Photo

🤖 Imagine a world where everyone has a personalized LLM agent by their side! Our latest research (arxiv.org/abs/2502.12149) explores how persona dynamics influence multi-agent competitive auctions. 🏆 Discover who wins, who loses, and the impact of personas on competitiveness!

🤖 Imagine a world where everyone has a personalized LLM agent by their side! Our latest research (arxiv.org/abs/2502.12149) explores how persona dynamics influence multi-agent competitive auctions. 🏆 Discover who wins, who loses, and the impact of personas on competitiveness!
Wenjie Jacky Mo (@wenjie_jacky_mo) 's Twitter Profile Photo

Worried about backdoors in LLMs? 🌟 Check out our #NAACL2025 work on test-time backdoor mitigation! ✅ Black-box 📦 ✅ Plug-and-play 🛡️ We explore: → Defensive Demonstrations 🧪 → Self-generated Prefixes 🧩 → Self-refinement ✍️ 📄 arxiv.org/abs/2311.09763 🧵[1/n]

Worried about backdoors in LLMs?

🌟 Check out our #NAACL2025 work on test-time backdoor mitigation!

✅ Black-box 📦
✅ Plug-and-play 🛡️

We explore:
→ Defensive Demonstrations 🧪
→ Self-generated Prefixes 🧩
→ Self-refinement ✍️

📄 arxiv.org/abs/2311.09763

🧵[1/n]
Tu Vu (@tuvllms) 's Twitter Profile Photo

📢 Research internship Google📢 I am looking for a PhD student researcher to work with me and my colleagues on advanced reasoning and/or RAG factuality this summer Google Mountain View, CA. We will focus on open-source models and benchmarks, and aim to publish our findings.

Fei Liu @ #ICLR2025 (@feiliu_nlp) 's Twitter Profile Photo

Curious how LLMs tackle planning tasks, such as travel and computer use? Our new survey #PlanGenLLMs (arxiv.org/abs/2502.11221) builds on classic work by Kartam and Wilkins (1990) and examines 6 key metrics to compare today's top planning systems. Your next agentic workflow

Curious how LLMs tackle planning tasks, such as travel and computer use? Our new survey #PlanGenLLMs (arxiv.org/abs/2502.11221) builds on classic work by Kartam and Wilkins (1990) and examines 6 key metrics to compare today's top planning systems. 

Your next agentic workflow
Yu Feng (@anniefeng6) 's Twitter Profile Photo

#ICLR2025 Oral LLMs often struggle with reliable and consistent decisions under uncertainty 😵‍💫 — largely because they can't reliably estimate the probability of each choice. We propose BIRD 🐦, a framework that significantly enhances LLM decision making under uncertainty. BIRD

#ICLR2025 Oral

LLMs often struggle with reliable and consistent decisions under uncertainty 😵‍💫 — largely because they can't reliably estimate the probability of each choice.

We propose BIRD 🐦, a framework that significantly enhances LLM decision making under uncertainty.

BIRD
Xiaogeng Liu (@xiaogengliu) 's Twitter Profile Photo

Thrilled to be featured in the #ICLR2025 Spotlight! 🎉 Come see our poster in Hall 3 + Hall 2B #602, April 25, 10:00–12:30 PM SGT

Cognitive Computation Group (@cogcomp) 's Twitter Profile Photo

Excited to share our papers at #ICLR2025 in Singapore! Check out the summaries on our blog (ccgblog.seas.upenn.edu/2025/04/ccg-pa…), and then check out the papers at oral session 1B (BIRD) and poster session 2 (for all three)! Yu Feng, Xingyu Fu, Ben Zhou, 🌴Muhao Chen🌴, Dan Roth

Excited to share our papers at #ICLR2025 in Singapore!  Check out the summaries on our blog (ccgblog.seas.upenn.edu/2025/04/ccg-pa…), and then check out the papers at oral session 1B (BIRD) and poster session 2 (for all three)!
<a href="/AnnieFeng6/">Yu Feng</a>, <a href="/XingyuFu2/">Xingyu Fu</a>, <a href="/BenZhou96/">Ben Zhou</a>, <a href="/muhao_chen/">🌴Muhao Chen🌴</a>, <a href="/DanRothNLP/">Dan Roth</a>
Fei Wang (@fwang_nlp) 's Twitter Profile Photo

🎉 Excited to share that our paper, "MuirBench: A Comprehensive Benchmark for Robust Multi-image Understanding", will be presented at #ICLR2025!​ 📅 Date: April 24 🕒 Time: 3:00 PM 📍 Location: Hall 3 + Hall 2B #11 MuirBench challenges multimodal LLMs with diverse multi-image

🎉 Excited to share that our paper, "MuirBench: A Comprehensive Benchmark for Robust Multi-image Understanding", will be presented at #ICLR2025!​
📅 Date: April 24
🕒 Time: 3:00 PM
📍 Location: Hall 3 + Hall 2B #11
MuirBench challenges multimodal LLMs with diverse multi-image
Nan Xu (@xunannancy) 's Twitter Profile Photo

Should multimodal in-context learning demonstrations mirror the visual or textual patterns of test instances? I'll present paper “From Introspection to Best Practices: Principled Analysis of Demonstrations in Multimodal In-Context Learning” on April 30th at NAACL. See you there!

Should multimodal in-context learning demonstrations mirror the visual or textual patterns of test instances?
I'll present paper “From Introspection to Best Practices: Principled Analysis of Demonstrations in Multimodal In-Context Learning” on April 30th at NAACL. 
See you there!
ComputerUseAgents Workshop (@workshopcua) 's Twitter Profile Photo

🔦 Speaker Spotlight: Sercan Ö. Arık (Sercan Arık) We're thrilled to welcome Sercan, Senior Staff Research Scientist at Google Cloud AI, as an invited speaker at the ICML 2025 Workshop on Computer Use Agents! Sercan's focuses on democratizing AI and applying it to impactful use

🔦 Speaker Spotlight: Sercan Ö. Arık (<a href="/sercanarik/">Sercan Arık</a>)
We're thrilled to welcome Sercan, Senior Staff Research Scientist at Google Cloud AI, as an invited speaker at the ICML 2025 Workshop on Computer Use Agents!
Sercan's focuses on democratizing AI and applying it to impactful use
Tianqing Fang @ ACL24 (@tfang229) 's Twitter Profile Photo

🚀 Check out our paper: WebEvolver: Enhancing Web Agent Self-Improvement with Coevolving World Model, from Tencent AI Lab!. We present a world model-driven framework for self-improving web agents, addressing critical challenges in self-training—such as limited exploration and

🚀 Check out our paper: WebEvolver: Enhancing Web Agent Self-Improvement with Coevolving World Model, from Tencent AI Lab!.

We present a world model-driven framework for self-improving web agents, addressing critical challenges in self-training—such as limited exploration and