Hung-yi Lee (李宏毅) (@hungyilee2) 's Twitter Profile
Hung-yi Lee (李宏毅)

@hungyilee2

Hung-yi Lee is currently a professor at National Taiwan University. He owns a YouTube channel teaching deep learning in Mandarin.

ID: 1239783572661100544

calendar_today17-03-2020 05:20:34

95 Tweet

4,4K Followers

19 Following

Hung-yi Lee (李宏毅) (@hungyilee2) 's Twitter Profile Photo

Launched "Intro to Generative AI" course with 1000+ students this spring! Thanks to Cheng Han Chiang (姜成翰) for being head TA. Using LLMs to evaluate assignments, inspired by his ACL paper (arxiv.org/abs/2305.01937). Check what we learned: arxiv.org/abs/2407.05216

Hung-yi Lee (李宏毅) (@hungyilee2) 's Twitter Profile Photo

SPS SLTC/AASP TC Webinar Don't miss out on recent advances in speech separation, end-to-end modeling, speaker diarization, and more! Speaker: Dr. Takuya Yoshioka, Director of Research at Assembly AI Inc. Time: 1:00 PM ET, 23 July 2024 Register here: landing.signalprocessingsociety.org/sltc-aasp

Hung-yi Lee (李宏毅) (@hungyilee2) 's Twitter Profile Photo

The paper "Self-Supervised Speech Representation Learning: A Review" is a top 25 download in IEEE JSTSP! The authors will discuss the latest in speech foundation models. Time: 1:00 PM ET, 6 Aug 2024 Registration page: signalprocessingsociety.org/blog/sps-webin…

The paper "Self-Supervised Speech Representation Learning: A Review" is a top 25 download in IEEE JSTSP! The authors will discuss the latest in speech foundation models.  

Time: 1:00 PM ET, 6 Aug 2024 
Registration page: signalprocessingsociety.org/blog/sps-webin…
Manling Li (@manlingli_) 's Twitter Profile Photo

Congrats to Cheng Han Chiang Cheng Han Chiang (姜成翰) and Hung-yi Lee Hung-yi Lee (李宏毅) for the Best Paper Award at ACL24 Knowledgeable LMs workshop! Merging Facts, Crafting Fallacies: Evaluating the Contradictory Nature of Aggregated Factual Claims in Long-Form Generations Cheng-Han Chiang, Hung-yi

Congrats to Cheng Han Chiang <a href="/dcml0714/">Cheng Han Chiang (姜成翰)</a> and Hung-yi Lee <a href="/HungyiLee2/">Hung-yi Lee (李宏毅)</a> for the Best Paper Award at ACL24 Knowledgeable LMs workshop!

Merging Facts, Crafting Fallacies: Evaluating the Contradictory Nature of Aggregated Factual Claims in Long-Form Generations
Cheng-Han Chiang, Hung-yi
Hung-yi Lee (李宏毅) (@hungyilee2) 's Twitter Profile Photo

Congratulations to Cheng Han Chiang (Cheng Han Chiang (姜成翰)) for winning the Best Paper Award at the ACL24 Knowledgeable LMs workshop! This paper tackles the issue I mentioned in my course (reurl.cc/LlmN2X) — combining correct facts can sometimes result in an incorrect response.

Hung-yi Lee (李宏毅) (@hungyilee2) 's Twitter Profile Photo

I'll give an overview talk on Spoken Language Models at INTERSPEECH 2024! Join me tomorrow, September 3rd, from 13:30 to 14:10 in the "Lasso" room. link of slides: drive.google.com/file/d/1gPjnjG…

Hung-yi Lee (李宏毅) (@hungyilee2) 's Twitter Profile Photo

Excited to speak at CHIME 2024, collocated with INTERSPEECH! Join me on Sept 6th, 14:00-15:00, for "Teaching New Skills to Foundation Models: Insights and Experiences." Learn why fine-tuning is more challenging than it seems! Workshop link: chimechallenge.org/current/worksh…

Hung-yi Lee (李宏毅) (@hungyilee2) 's Twitter Profile Photo

Congratulations to the SUPERB Team! Our work on the Speech Processing Universal PERformance Benchmark (SUPERB) has been ranked 7th among the most cited papers at INTERSPEECH over the past five years! A big round of applause to everyone involved.

Congratulations to the SUPERB Team! Our work on the Speech Processing Universal PERformance Benchmark (SUPERB) has been ranked 7th among the most cited papers at INTERSPEECH over the past five years! A big round of applause to everyone involved.
Hung-yi Lee (李宏毅) (@hungyilee2) 's Twitter Profile Photo

🚀 Excited to announce that DeSTA2, a spoken LLM. What's amazing? It's trained on just 150 hours of speech data, beating models with tens of thousands of hours! Shoutout to NTU PhD student Ke-Han Lu & NVIDIA team for leading this! 👉 Learn more: kehanlu.github.io/DeSTA2/

🚀 Excited to announce that DeSTA2, a spoken LLM. What's amazing? It's trained on just 150 hours of speech data, beating models with tens of thousands of hours! Shoutout to NTU PhD student Ke-Han Lu &amp; NVIDIA team for leading this!

👉 Learn more: kehanlu.github.io/DeSTA2/
Hung-yi Lee (李宏毅) (@hungyilee2) 's Twitter Profile Photo

Check out this new paper using RLAIF to enhance Spoken LLM! Research by NTU PhD student Guan-Ting Lin and Amazon researchers. link: arxiv.org/abs/2411.01834

Check out this new paper using RLAIF to enhance Spoken LLM! Research by NTU PhD student Guan-Ting Lin and Amazon researchers.

link: arxiv.org/abs/2411.01834
Hung-yi Lee (李宏毅) (@hungyilee2) 's Twitter Profile Photo

I'll be attending the IEEE Spoken Language Technology Workshop (SLT 2024) in Macao from December 2nd to 5th! Looking forward to engaging with fellow researchers and exploring the latest advancements in speech and language processing. 2024.ieeeslt.org

I'll be attending the IEEE Spoken Language Technology Workshop (SLT 2024) in Macao from December 2nd to 5th! Looking forward to engaging with fellow researchers and exploring the latest advancements in speech and language processing.

2024.ieeeslt.org
Hung-yi Lee (李宏毅) (@hungyilee2) 's Twitter Profile Photo

Excited about neural audio codecs and speech LLMs? Don’t miss the Codec-SUPERB @ SLT24 special session, on December 3rd, 15:00–18:30 in Macau! We’re thrilled to host prominent speakers from cutting-edge projects, including SemantiCodec, BNN, Uniaudio, VoiceCraft, and Moshi.

Excited about neural audio codecs and speech LLMs? Don’t miss the Codec-SUPERB @ SLT24 special session, on December 3rd, 15:00–18:30 in Macau!

We’re thrilled to host prominent speakers from cutting-edge projects, including SemantiCodec, BNN, Uniaudio, VoiceCraft, and Moshi.
Shinji Watanabe (@shinjiw_at_cmu) 's Twitter Profile Photo

We are excited to announce the launch of ML SUPERB 2.0 (multilingual.superbbenchmark.org) as part of the Interspeech 2024 official challenge! We hope this upgraded version of ML SUPERB advances universal access to speech processing worldwide. Please join it! #Interspeech2025

We are excited to announce the launch of ML SUPERB 2.0 (multilingual.superbbenchmark.org) as part of the Interspeech 2024 official challenge! We hope this upgraded version of ML SUPERB advances universal access to speech processing worldwide. Please join it! 
#Interspeech2025
Hung-yi Lee (李宏毅) (@hungyilee2) 's Twitter Profile Photo

I will deliver the final talk in the SPS SLTC/AASP TC Webinar Series in 2024, sharing insights on fine-tuning models such as LLaMA and Whisper. 📅 Dec 17, 10 AM ET (11 PM Taiwan). Register: signalprocessingsociety.org/blog/sps-sltca…

Xuanjun (Victor) Chen 🤖 (@xjchen_ntu) 's Twitter Profile Photo

🚨 Call for Papers – ASRU 2025 Special Session 🎤 Responsible Speech & Audio Generative AI 📍 Honolulu, Hawaii · Dec 2025 Join us to tackle accountability, fairness, and trust in generative speech/music/audio systems! 👉 Deadline: May 28, 2025 🔗 Detail: codecfake.github.io/RespSA-GenAI/

🚨 Call for Papers – ASRU 2025 Special Session
🎤 Responsible Speech &amp; Audio Generative AI
📍 Honolulu, Hawaii · Dec 2025
Join us to tackle accountability, fairness, and trust in generative speech/music/audio systems!
👉 Deadline: May 28, 2025
🔗 Detail: codecfake.github.io/RespSA-GenAI/
Hung-yi Lee (李宏毅) (@hungyilee2) 's Twitter Profile Photo

🎙️ Upcoming IEEE SPS SLTC/AASP Webinar! 📅 May 8, 2025 | 🕗 8:00 AM ET 🔹 Topic: Audio Signal Processing with Microphone Arrays 🎤 Speaker: Dr. Sharon Gannot Register now 👉 landing.signalprocessingsociety.org/may-08-2025

Hung-yi Lee (李宏毅) (@hungyilee2) 's Twitter Profile Photo

🎙️ Introducing TASTE: a novel audio tokenization approach for speech LLMs, by Liang-Hsuan Tseng (NTU), Yi-Chang Chen (MTK Research), and Kuan-Yi Lee (NTU)! 🔗 Paper: arxiv.org/abs/2504.07053 📺 Mandarin introduction (from 1:03:30): youtu.be/gkAyqoQkOSk?si…

🎙️ Introducing TASTE: a novel audio tokenization approach for speech LLMs, by Liang-Hsuan Tseng (NTU), Yi-Chang Chen (MTK Research), and Kuan-Yi Lee (NTU)!

🔗 Paper: arxiv.org/abs/2504.07053

📺 Mandarin introduction (from 1:03:30): youtu.be/gkAyqoQkOSk?si…
Hung-yi Lee (李宏毅) (@hungyilee2) 's Twitter Profile Photo

🎧 With the rapid growth of audio LLM benchmarking studies, a comprehensive survey is timely! Check out the survey paper on benchmarks in audio LLMs by Chih-Kai Yang and Neo S. Ho. 🔥 Paper link: arxiv.org/abs/2505.15957

Pooneh Mousavi (@mousavipooneh) 's Twitter Profile Photo

🚀 We're excited to announce our latest work: "Discrete Audio Tokens: More Than a Survey!" It presents a comprehensive survey and benchmark of audio tokenizers across speech, music, and general audio. preprint: arxiv.org/pdf/2506.10274 website: poonehmousavi.github.io/dates-website/

Hung-yi Lee (李宏毅) (@hungyilee2) 's Twitter Profile Photo

📢 Deadline Extended! Submit your original research papers to the conference track of the Asian Conference on Machine Learning (ACML 2025) by June 26, 2025! Join ACML 2025 in Taipei, Taiwan (Dec 9–12). Details: acml-conf.org/2025/index.htm