Yingheng Wang (@yingheng_wang) 's Twitter Profile
Yingheng Wang

@yingheng_wang

CS PhD Student @Cornell @CUAISci | Intern @awscloud @MSFTResearch | Prev @JohnsHopkins, @Tsinghua_Uni | ML & AI4Science

ID: 1478466316088619011

linkhttp://isjakewong.github.io calendar_today04-01-2022 20:40:24

215 Tweet

769 Followers

399 Following

Yingheng Wang (@yingheng_wang) 's Twitter Profile Photo

Wenting is an amazing and talented researcher and really into the reasoning world. She's the force behind WildChat and Commit 0. If you're into top-notch research, you should definitely consider working with her!

Kexin Huang (@kexinhuang5) 's Twitter Profile Photo

📢 Introducing Biomni - the first general-purpose biomedical AI agent. Biomni is built on the first unified environment for biomedical agent with 150 tools, 59 databases, and 106 software packages and a generalist agent design with retrieval, planning, and code as action. This

Albert Tseng (@tsengalb99) 's Twitter Profile Photo

📣Introducing our latest work: Yet Another Quantization Algorithm! YAQA directly minimizes the KL divergence to the original model during rounding, cutting it by >30% over prior PTQ methods and giving an even closer model than Google’s QAT on Gemma! 🤯 arxiv.org/abs/2505.22988👇

📣Introducing our latest work: Yet Another Quantization Algorithm!

YAQA directly minimizes the KL divergence to the original model during rounding, cutting it by >30% over prior PTQ methods and giving an even closer model than Google’s QAT on Gemma! 🤯

arxiv.org/abs/2505.22988👇
Andrej Karpathy (@karpathy) 's Twitter Profile Photo

My sleep scores during recent travel were in the 90s. Now back in SF I am consistently back down to 70s, 80s. I am increasingly convinced that this is due to traffic noise from a nearby road/intersection where I live - every ~10min, a car, truck, bus, or motorcycle with a very

Xinyu Yang (@xinyu2ml) 's Twitter Profile Photo

🚀 Super excited to share Multiverse! 🏃 It’s been a long journey exploring the space between model design and hardware efficiency. What excites me most is realizing that, beyond optimizing existing models, we can discover better model architectures by embracing system-level

Yingheng Wang (@yingheng_wang) 's Twitter Profile Photo

I would say TMLR offers a better submission experience than mainstream ML conferences nowadays: transparent and responsible Action Editors, reviewers who focus on genuinely improving your paper (e.g., deeper discussion), not vague "novelty" or endless baselines. If you adequately

I would say TMLR offers a better submission experience than mainstream ML conferences nowadays: transparent and responsible Action Editors, reviewers who focus on genuinely improving your paper (e.g., deeper discussion), not vague "novelty" or endless baselines. If you adequately
Xiuyu Li (@xiuyu_l) 's Twitter Profile Photo

Sparsity can make your LoRA fine-tuning go brrr 💨 Announcing SparseLoRA (ICML 2025): up to 1.6-1.9x faster LLM fine-tuning (2.2x less FLOPs) via contextual sparsity, while maintaining performance on tasks like math, coding, chat, and ARC-AGI 🤯 🧵1/ z-lab.ai/projects/spars…