Center for AI Safety (@ai_risks) 's Twitter Profile
Center for AI Safety

@ai_risks

Reducing societal-scale risks from AI.
safe.ai
ai-frontiers.org

ID: 1562139162781704192

linkhttps://safe.ai calendar_today23-08-2022 18:06:33

198 Tweet

6,6K Followers

3 Following

AI Frontiers (@aif_media) 's Twitter Profile Photo

Does an AI rivalry between the US and China actually make us safer? Two AI policy researchers argue that Washington must pivot from power plays to partnership, working with Beijing to forge joint safety guardrails and avert catastrophic AI risks: ai-frontiers.org/articles/ameri…

AI Frontiers (@aif_media) 's Twitter Profile Photo

Can the US Prevent AGI from Being Stolen? Securing AI weights from foreign adversaries would require a level of security never seen before. We explore what this would take. ai-frontiers.org/articles/can-t…

Center for AI Safety (@ai_risks) 's Twitter Profile Photo

"Dynamism vs. stasis" is a clearer lens for AI safety debates. Helen Toner (Helen Toner) argues that many AI safety ideas lean too far toward control and rigidity—threatening a dynamic, open-ended future. Read it on AI Frontiers: ai-frontiers.org/articles/were-…

Center for AI Safety (@ai_risks) 's Twitter Profile Photo

Interpretability research aims to reverse-engineer AI, yet despite a decade of effort, Dan Hendrycks argues progress has been minimal. Complex AI systems may simply defy a neat, neuron-by-neuron explanation, raising questions about the future of 'mechanistic interpretability.

Interpretability research aims to reverse-engineer AI, yet despite a decade of effort, <a href="/DanHendrycks/">Dan Hendrycks</a>  argues progress has been minimal. Complex AI systems may simply defy a neat, neuron-by-neuron explanation, raising questions about the future of 'mechanistic interpretability.
Center for AI Safety (@ai_risks) 's Twitter Profile Photo

Applications are open for the Summer (Jun 23–Sep 14) session of our free, online AI Safety, Ethics, and Society course! The course is open to all fields—no technical background required. Apply by May 30. More info & application: aisafetybook.com/virtual-course

Center for AI Safety (@ai_risks) 's Twitter Profile Photo

Up to 50% of China's AI training is run on smuggled U.S. chips. Why? The US can't track chips after export. Congress just proposed a solution:  location verification. We break down how it could work: ai-frontiers.org/articles/locat…

Center for AI Safety (@ai_risks) 's Twitter Profile Photo

Applications are open for the Winter (Nov 3–Feb 1) session of our free, online AI Safety, Ethics, and Society course! Open to all fields—no technical background required. Apply by October 10. More info & apply: aisafetybook.com/virtual-course

Dan Hendrycks (@danhendrycks) 's Twitter Profile Photo

The term “AGI” is currently a vague, moving goalpost. To ground the discussion, we propose a comprehensive, testable definition of AGI. Using it, we can quantify progress: GPT-4 (2023) was 27% of the way to AGI. GPT-5 (2025) is 58%. Here’s how we define and measure it: 🧵

The term “AGI” is currently a vague, moving goalpost.

To ground the discussion, we propose a comprehensive, testable definition of AGI.
Using it, we can quantify progress:
GPT-4 (2023) was 27% of the way to AGI. GPT-5 (2025) is 58%.

Here’s how we define and measure it: 🧵