Chi Han (@glaciohound) 's Twitter Profile
Chi Han

@glaciohound

CS PhD student at UIUC, interested in language models and their understanding.

ID: 1438485864393314311

linkhttps://glaciohound.github.io/ calendar_today16-09-2021 12:52:30

54 Tweet

745 Followers

264 Following

Manling Li (@manlingli_) 's Twitter Profile Photo

We got the #ACL2024 Outstanding Paper for “LM-Steer: Word Embeddings Are Steers for Language Models”! A big shoutout and congrats to our amazing leader Chi Han Chi Han and Heng Ji, and to our wonderful team Jialiang Xu @YiFung10 Chenkai Sun Nan Jiang Tarek!

Chi Han (@glaciohound) 's Twitter Profile Photo

🎖Excited that "LM-Steer: Word Embeddings Are Steers for Language Models" became my another 1st-authored Outstanding Paper #ACL2024 (besides LM-Infinite) We revealed steering roles of word embeddings for continuous, compositional, efficient, interpretable& transferrable control!

🎖Excited that "LM-Steer: Word Embeddings Are Steers for Language Models" became my another 1st-authored Outstanding Paper #ACL2024 (besides LM-Infinite)

We revealed steering roles of word embeddings for continuous, compositional, efficient, interpretable& transferrable control!
Dylan (@dylan_works_) 's Twitter Profile Photo

"Only-IF: Revealing the Decisive Effect of Instruction Diversity on Generalization" arxiv.org/pdf/2410.04717 We isolated 'instruction-following' ability (apart from complex reasoning like math) and designed various controlled experiments to show that -

Ke Yang (@empathyang) 's Twitter Profile Photo

👾 Introducing AgentOccam: Automating Web Tasks with LLMs! 🌐 AgentOccam showcases the impressive power of Large Language Models (LLMs) on web tasks, without any in-context examples, new agent roles, online feedback, or search strategies. 🏄🏄🏄 🧙 Link: arxiv.org/abs/2410.13825

👾 Introducing AgentOccam: Automating Web Tasks with LLMs! 🌐 AgentOccam showcases the impressive power of Large Language Models (LLMs) on web tasks, without any in-context examples, new agent roles, online feedback, or search strategies. 🏄🏄🏄
🧙 Link: arxiv.org/abs/2410.13825
Jiaxin-Qin (@jr_qjx) 's Twitter Profile Photo

I am at #EMNLP2024! I will present our work "Why Does New Knowledge Create Messy Ripple Effects in LLMs? " on Web 10:30am. Thanks to all the collaborators Heng Ji Zixuan Zhang Chi Han Manling Li Looking forward to have a chat! Paper Link: arxiv.org/pdf/2407.12828

I am at #EMNLP2024!

I will present our work "Why Does New Knowledge Create Messy Ripple Effects in LLMs? " on Web 10:30am. 

Thanks to all the collaborators <a href="/hengjinlp/">Heng Ji</a> <a href="/zhangzxUIUC/">Zixuan Zhang</a> <a href="/Glaciohound/">Chi Han</a> <a href="/ManlingLi_/">Manling Li</a> 

Looking forward to have a chat! 

Paper Link: arxiv.org/pdf/2407.12828
Rui Pan (@rui4research) 's Twitter Profile Photo

Presenting our LISA paper at NeurIPS 2024😆 - Dec. 13 at 4:30 pm (Friday afternoon) - West Ballroom A-D #5708 Fine-tuning 7B in a single GPU❔ Randomly freezing ~90% self-attention layers every 5-20 iterations allows that!🚀It is - 3x Fast - Memory-efficient - Good at

Presenting our LISA paper at NeurIPS 2024😆
- Dec. 13 at 4:30 pm (Friday afternoon)
- West Ballroom A-D #5708

Fine-tuning 7B in a single GPU❔ Randomly freezing ~90% self-attention layers every 5-20 iterations allows that!🚀It is
- 3x Fast
- Memory-efficient
- Good at
Qingyun Wang (@eagle_hz) 's Twitter Profile Photo

📢📈 I’m on the 2025 faculty job market! I've been incredibly grateful to work with inspiring advisors, mentors & peers. 💡My research, AI4Scientists🔬, accelerates & democratizes the research lifecycle by: 1️⃣ Few-shot scientific knowledge acquisition 2️⃣ Domain-aware scientific

📢📈 I’m on the 2025 faculty job market!
I've been incredibly grateful to work with inspiring advisors, mentors &amp; peers.
💡My research, AI4Scientists🔬, accelerates &amp; democratizes the research lifecycle by:
1️⃣ Few-shot scientific knowledge acquisition
2️⃣ Domain-aware scientific
Ke Yang (@empathyang) 's Twitter Profile Photo

👽New release: We filter out educationally valuable web data rather than using arXiv papers to continually pre-train a specialist Astro LLM. Big thanks to the first author Eric Modesitt for the project leadership! 💪 Great ideas and strong execution—you're an amazing undergrad!

Chi Han (@glaciohound) 's Twitter Profile Photo

Thank you so much, Heng Heng Ji! I'm incredibly grateful for your guidance and support throughout this journey and honored to receive the fellowship to continue working on exciting projects. Looking forward to more collaborations with you and Avi Sil Avi Sil @ ACL 2025 (virtually)! 🙌

Ke Yang (@empathyang) 's Twitter Profile Photo

🙌 Happy New Year everyone! 🤖 New preprint: TinyHelen's First Curriculum: Training and Evaluating Tiny Language Models in a Simpler Language Environment 🤖 We train and evaluate tiny language models (LMs) using a novel text dataset with systematically simplified vocabularies and

🙌 Happy New Year everyone!
🤖 New preprint: TinyHelen's First Curriculum: Training and Evaluating Tiny Language Models in a Simpler Language Environment
🤖 We train and evaluate tiny language models (LMs) using a novel text dataset with systematically simplified vocabularies and
Yu Wang (@__yuwang__) 's Twitter Profile Photo

(1/4) Excited to share that our position paper "Towards LifeSpan Cognitive Systems" has been accepted to TMLR! As "agents" dominate the AI landscape in 2025, we push the boundaries by envisioning LSCS — AI systems that go beyond personal assistants to lifespan cognitive systems.

(1/4) Excited to share that our position paper "Towards LifeSpan Cognitive Systems" has been accepted to TMLR! As "agents" dominate the AI landscape in 2025, we push the boundaries by envisioning LSCS — AI systems that go beyond personal assistants to lifespan cognitive systems.
Dylan (@dylan_works_) 's Twitter Profile Photo

Sharing our recent study - "The Best Instruction-Tuning Data are Those That Fit". We introduce a practical, lightweight approach that selects in-distribution data for supervised fine-tuning—yielding better models with less data & compute w. surprising simplicity and efficiency!

Chi Han (@glaciohound) 's Twitter Profile Photo

Welcome to my #AAAI2025 Tutorial, "The Quest for A Science of LMs," today! Time: Feb 26, 2pm-3:45pm Location: Room 113A, Pennsylvania Convention Center Website: glaciohound.github.io/Science-of-LLM… Underline: underline.io/events/487/sch…

Chi Han (@glaciohound) 's Twitter Profile Photo

Thanks, Heng Ji, for sharing our recent work! In this paper, we investigate how LLMs achieve positional flexibility, such as understanding shuffled sentences and reading long context (a phenomenon also observed in humans). This paper uncovers the hidden computational

Chi Han (@glaciohound) 's Twitter Profile Photo

New Benchmark for LLM Multi-Turn-Instruct Following!🚀 Can LLMs handle multiple entangled instructions effectively? Our comprehensive Multi-Turn-Instruct dataset evaluates LLMs’ ability to 📜track, 🔄reason on, and ⚖️ resolve conflicts across multiple turns of instructions in