Haishuo (@haishuofang) 's Twitter Profile
Haishuo

@haishuofang

NLPer, PhD candidate at UKP Lab

ID: 850350298715443201

calendar_today07-04-2017 14:11:24

7 Tweet

79 Followers

235 Following

Nafise Sadat Moosavi (@nafisesadat) 's Twitter Profile Photo

Activation functions reduce the topological complexity of data. Best AF may be diff for diff models and diff layers, but most Transformer models use GELU. What if the model learns optimized activation functions during training? led by Haishuo with Ji Ung Lee and Iryna Gurevych

Haritz Puerto @ NAACL 2025 🌵🇺🇸 (@haritzpuerto) 's Twitter Profile Photo

Super excited to have a second paper at #ACL2023NLP !!🎉 In this paper, we propose a new method to teach NLP in a more interactive way using SQuARE Project Super happy to contribute to improving NLP teaching :)

UKP Lab (@ukplab) 's Twitter Profile Photo

Are Emergent Abilities in Large Language Models just In-Context Learning? Spoiler: YES 🤯 Through a series of over 1,000 experiments, we provide compelling evidence: arxiv.org/abs/2309.01809 Our results allay safety concerns regarding latent hazardous abilities. A🧵👇 #NLProc

Are Emergent Abilities in Large Language Models just In-Context Learning?

Spoiler: YES 🤯

Through a series of over 1,000 experiments, we provide compelling evidence: arxiv.org/abs/2309.01809 

Our results allay safety concerns regarding latent hazardous abilities.

A🧵👇 #NLProc