Xiaozhi Wang (@xiaozhiwangnlp) 's Twitter Profile
Xiaozhi Wang

@xiaozhiwangnlp

Ph.D. candidate @Tsinghua_Uni | NLP/KG

ID: 933539856331841536

linkhttps://bakser.github.io/ calendar_today23-11-2017 03:37:00

22 Tweet

210 Followers

408 Following

Jian Tang (@tangjianpku) 's Twitter Profile Photo

I recently gave a tea talk on graph representation learning at milamontreal, which summarized some of the most important work my group have done . I started to work on this back to 2013 and published the first paper LINE. Slides are available at: github.com/tangjianpku/ta…

I recently gave a tea talk on graph representation learning at <a href="/MILAMontreal/">milamontreal</a>, which summarized some of the most important work my group have done . I started to work on this back to 2013 and published the first paper LINE. Slides are available at: github.com/tangjianpku/ta…
TsinghuaNLP (@tsinghuanlp) 's Twitter Profile Photo

Welcome to the TsinghuaNLP Twitter feed, where we'll share new researches and information from TsinghuaNLP Group. Looking forward to interacting with you here! #NLProc

Welcome to the <a href="/TsinghuaNLP/">TsinghuaNLP</a> Twitter feed, where we'll share new researches and information from TsinghuaNLP Group. Looking forward to interacting with you here! #NLProc
TsinghuaNLP (@tsinghuanlp) 's Twitter Profile Photo

KEPLER (arxiv.org/pdf/1911.06136…) is a unified model for knowledge embedding (KE) and pre-trained language model (PLM) representation. It encodes textual entity descriptions with a PLM and then jointly optimize KE and language modeling objectives. #NLProc #AI #TsinghuaNLP

KEPLER (arxiv.org/pdf/1911.06136…) is a unified model for knowledge embedding (KE) and pre-trained language model (PLM) representation. It encodes textual entity descriptions with a PLM and then jointly optimize KE and language modeling objectives. #NLProc #AI #TsinghuaNLP
TsinghuaNLP (@tsinghuanlp) 's Twitter Profile Photo

MAVEN is a massive general domain event detection dataset, which contains 4,480 Wikipedia documents, 118,732 event mentions, and 168 event types. Check out our EMNLP 2020 paper (arxiv.org/abs/2004.13590). #NLProc #AI #TsinghuaNLP #EMNLP

MAVEN is a massive general domain event detection dataset, which contains 4,480 Wikipedia documents, 118,732 event mentions, and 168 event types. Check out our EMNLP 2020 paper (arxiv.org/abs/2004.13590). #NLProc #AI #TsinghuaNLP #EMNLP
TsinghuaNLP (@tsinghuanlp) 's Twitter Profile Photo

HMEAE is an event argument extraction model. It provides effective inductive bias from the concept hierarchy of event argument roles by using a neural module network to imitate the hierarchical structure of concepts. #NLProc #AI #TsinghuaNLP #EMNLP Paper: aclweb.org/anthology/D19-…

HMEAE is an event argument extraction model. It provides effective inductive bias from the concept hierarchy of event argument roles by using a neural module network to imitate the hierarchical structure of concepts. #NLProc #AI #TsinghuaNLP #EMNLP
Paper: aclweb.org/anthology/D19-…
TsinghuaNLP (@tsinghuanlp) 's Twitter Profile Photo

Our ACL 2021 paper proposed a contrastive pre-training framework CLEVE for event extraction task, utilizing AMR structures to make PLMs better understand semantic relations between triggers and arguments. Our paper: arxiv.org/abs/2105.14485. #NLProc #ACL2021 #AI #TsinghuaNLP

Our ACL 2021 paper proposed a contrastive pre-training framework CLEVE for event extraction task, utilizing AMR structures to make PLMs better understand semantic relations between triggers and arguments. Our paper: arxiv.org/abs/2105.14485. #NLProc #ACL2021 #AI #TsinghuaNLP
Yujia Qin@ICLR2025 (@tsingyoga) 's Twitter Profile Photo

Wanna tune few-shot NLP tasks with 5 free parameters😉? Check our recent preprint: Exploring Low-dimensional Intrinsic Task Subspace via Prompt Tuning. We explore how can PLMs effectively adapt to broad NLP tasks differing a lot superficially. arxiv.org/abs/2110.07867

TsinghuaNLP (@tsinghuanlp) 's Twitter Profile Photo

🎉Thrilled to introduce our latest work, Delta Tuning: A Comprehensive Study of Parameter Efficient Methods for Pre-trained Language Models. We perform a comprehensive theoretical analysis and experimental validation of the parametrically efficient paradigm for #PLMs.

🎉Thrilled to introduce our latest work, Delta Tuning: A Comprehensive Study of Parameter Efficient Methods for Pre-trained Language Models. We perform a comprehensive theoretical analysis and experimental validation of the parametrically efficient paradigm for #PLMs.
TsinghuaNLP (@tsinghuanlp) 's Twitter Profile Photo

Welcome to follow our #NAACL-HLT 2022 paper. We empirically investigate the transferability of soft prompts across tasks/PLMs and introduce a new transferability indicator. #NLProc #AI #TsinghuaNLP (1/4) Paper:aclanthology.org/2022.naacl-mai… Data/Code: github.com/thunlp/Prompt-…

Welcome to follow our #NAACL-HLT 2022 paper. We empirically investigate the transferability of soft prompts across tasks/PLMs and introduce a new transferability indicator. #NLProc #AI #TsinghuaNLP (1/4)
Paper:aclanthology.org/2022.naacl-mai…
Data/Code: github.com/thunlp/Prompt-…
Chaojun Xiao (@xcjthu1) 's Twitter Profile Photo

1/5 🚀 Excited to share our latest paper on Configurable Foundation Models! 🧠 Inspired by the human brain's functional specialization, we propose a concept: Configurable Foundation Model, a modular approach to LLMs.

1/5 🚀 Excited to share our latest paper on Configurable Foundation Models! 🧠

Inspired by the human brain's functional specialization, we propose a concept: Configurable Foundation Model, a modular approach to LLMs.
Chaojun Xiao (@xcjthu1) 's Twitter Profile Photo

1/4 🚀 Densing Law of LLMs 🚀 OpenAI's Scaling Law showed how model capabilities scale with size. But what about the trend toward efficient models? 🤔 We introduce "capacity density" and found an exciting empirical law: LLMs' capacity density grows EXPONENTIALLY over time!

1/4 🚀 Densing Law of LLMs 🚀

OpenAI's Scaling Law showed how model capabilities scale with size. But what about the trend toward efficient models? 🤔

We introduce "capacity density" and found an exciting empirical law: LLMs' capacity density grows EXPONENTIALLY over time!
OpenBMB (@openbmb) 's Twitter Profile Photo

🚀 MiniCPM4 is here! 5x faster on end devices 🔥 ✨ What's new: 🏗️ Efficient Model Architecture - InfLLM v2 -- Trainable Sparse Attention Mechanism 🧠 Efficient Learning Algorithms - Model Wind Tunnel 2.0 -- Efficient Predictable Scaling - BitCPM -- Ultimate Ternary Quantization