Silviu Pitis (@silviupitis) 's Twitter Profile
Silviu Pitis

@silviupitis

ML PhD student at @UofT/@VectorInst working on normative AI alignment.

ID: 718091060786946049

linkhttps://silviupitis.com calendar_today07-04-2016 15:00:25

93 Tweet

2,2K Followers

731 Following

Keiran Paster (@keirp1) 's Twitter Profile Photo

Introducing OpenWebMath, a massive dataset containing every math document found on the internet - with equations in LaTeX format! 🤗 Download on @HuggingFace: huggingface.co/datasets/open-… 📝 Read the paper: arxiv.org/abs/2310.06786 w/ Marco Dos Santos, Zhangir Azerbayev, Jimmy Ba!

Introducing OpenWebMath, a massive dataset containing every math document found on the internet - with equations in LaTeX format!

🤗 Download on @HuggingFace: huggingface.co/datasets/open-…
📝 Read the paper: arxiv.org/abs/2310.06786

w/ <a href="/dsantosmarco/">Marco Dos Santos</a>, <a href="/zhangir_azerbay/">Zhangir Azerbayev</a>, <a href="/jimmybajimmyba/">Jimmy Ba</a>!
Yongchao Zhou (@yongchao_zhou_) 's Twitter Profile Photo

🎉 Excited to introduce DistillSpec! Accelerate your LLM inference using Speculative Decoding with a more aligned draft model, consistently delivering a remarkable 10-45% performance uplift over standard Speculative Decoding. ✨ Combining Distilled Target with Distilled Draft,

🎉 Excited to introduce DistillSpec! Accelerate your LLM inference using Speculative Decoding with a more aligned draft model, consistently delivering a remarkable 10-45% performance uplift over standard Speculative Decoding.
✨ Combining Distilled Target with Distilled Draft,
Ziang Xiao (@ziangxiao) 's Twitter Profile Photo

What constitutes a good #eval metric? We bridge #MeasurementTheory in #EducationalTesting and #Psychometrics to #NLProc. We propose #MetricEval, a theory-driven framework to conceptualize and operationalize measurement error sources to evaluate #NLG metrics. #EMNLP2023 🧵

What constitutes a good #eval metric? We bridge #MeasurementTheory in #EducationalTesting and #Psychometrics to #NLProc. We propose #MetricEval, a theory-driven framework to conceptualize and operationalize measurement error sources to evaluate #NLG metrics. #EMNLP2023 🧵
Rachel Freedman (@freedmanrach) 's Twitter Profile Photo

RLHF typically assumes that all training feedback comes from a single teacher, but teachers can disagree up to 37% of the time in practice. In our new paper, we introduce active teacher selection to learn from different teachers. (1/n)

Jiahai Feng (@feng_jiahai) 's Twitter Profile Photo

When given context about a “green square” and a “blue circle”, how do language models bind corresponding shapes and colors? Using causal experiments, we find that large enough language models learn simple structured representations for binding! A thread (1/n)

When given context about a “green square” and a “blue circle”, how do language models bind corresponding shapes and colors?

Using causal experiments, we find that large enough language models learn simple structured representations for binding!

A thread (1/n)
Alan Chan (@_achan96_) 's Twitter Profile Photo

OpenAI just announced GPTs and the Assistants API for “ helping developers build agent-like experiences”, but what does that mean and how does it change how we should govern AI? Some early thoughts relating to my ongoing work 🧵:

Yangjun Ruan (@yangjunr) 's Twitter Profile Photo

#OpenAI’s GPTs & Assistants APIs are a blast, making it much easier to build customized agents with new tools. But are they safe to deploy? 🚨 A simple & quick test against prompt injections reveals that it is fairly easy to make GPTs delete all your files 💀

#OpenAI’s GPTs &amp; Assistants APIs are a blast, making it much easier to build customized agents with new tools. But are they safe to deploy? 🚨

A simple &amp; quick test against prompt injections reveals that it is fairly easy to make GPTs delete all your files 💀
Silviu Pitis (@silviupitis) 's Twitter Profile Photo

I will be at #NeurIPS2023 Dec 11-16 Shoot me an email to connect! Particularly interested in: - LM eval for long-horizon / agents - Alignment / rewards generally Will present my paper on multi-objective reward aggregation at Poster sess 6 Thurs eve (neurips.cc/virtual/2023/p…)

Lucas Caccia (@lucaspcaccia) 's Twitter Profile Photo

Our team at MSR Montréal is looking for interns! Subjects range from efficient modular adaptation to building complex systems by stacking LLMs. Consider applying here : aka.ms/AAo5t0x

Yangjun Ruan (@yangjunr) 's Twitter Profile Photo

ToolEmu has been accepted at #ICLR2024 as a Spotlight presentation🔥 Explore our LLM-based emulation framework for identifying LLM agent risks at scale! 🎯 Demo: demo.toolemu.com 📄 Paper: arxiv.org/abs/2309.15817 🔗 Code: github.com/ryoungj/ToolEmu 🧵⬇️

Roger Grosse (@rogergrosse) 's Twitter Profile Photo

Here's what I see as a likely AGI trajectory over the next decade. I claim that later parts of the path present the biggest alignment risks/challenges. The alignment world has been focusing a lot on the lower left corner lately, which I'm worried is somewhat of a Maginot line.

Here's what I see as a likely AGI trajectory over the next decade.

I claim that later parts of the path present the biggest alignment risks/challenges. The alignment world has been focusing a lot on the lower left corner lately, which I'm worried is somewhat of a Maginot line.
Yangjun Ruan (@yangjunr) 's Twitter Profile Photo

We are presenting ToolEmu at #ICLR2024 tomorrow! ⏲️ Friday 4:30pm-6:30pm CEST 📍 Spotlight poster session, Hall B #80 I won't be able to attend ICLR this year but don't miss the chance to meet our amazing collaborators!

Blair Yang (@blairyang12) 's Twitter Profile Photo

🔍 Current LLM evaluations fall short: • Lack nuanced understanding of model capabilities • Overly focused on quantitative metrics • Difficult for humans to interpret Introducing LLM Report Cards! A novel approach for qualitative, interpretable model evaluation. 1/N

🔍 Current LLM evaluations fall short:
• Lack nuanced understanding of model capabilities
• Overly focused on quantitative metrics
• Difficult for humans to interpret
Introducing LLM Report Cards! A novel approach for qualitative, interpretable model evaluation.
1/N
Michael Zhang (@michaelrzhang) 's Twitter Profile Photo

📝 How do you choose which language model to use? Quantitative benchmarks can be uninformative and fall prey to Goodhart's Law, and even Chatbot Arena performance can be optimized for. In our new preprint, we propose generating qualitative report cards... 🧵

📝 How do you choose which language model to use? Quantitative benchmarks can be uninformative and fall prey to Goodhart's Law, and even Chatbot Arena performance can be optimized for.

In our new preprint, we propose generating qualitative report cards... 🧵
Schwartz Reisman Institute (@torontosri) 's Twitter Profile Photo

“What objective function do we want AI to optimize for? If we aggregate values from society, what weights do we use, and whose values?” Learn more about SRI Grad Affiliate Silviu Pitis's research, supported by an OpenAI Superalignment Fast Grant. 🔗 uoft.me/aWX

“What objective function do we want AI to optimize for? If we aggregate values from society, what weights do we use, and whose values?”

Learn more about SRI Grad Affiliate <a href="/silviupitis/">Silviu Pitis</a>'s research, supported by an <a href="/OpenAI/">OpenAI</a> Superalignment Fast Grant.

🔗 uoft.me/aWX