Center for AI Safety
@ai_risks
Reducing societal-scale risks from AI.
safe.ai
ai-frontiers.org
ID: 1562139162781704192
https://safe.ai 23-08-2022 18:06:33
198 Tweet
6,6K Followers
3 Following
"Dynamism vs. stasis" is a clearer lens for AI safety debates. Helen Toner (Helen Toner) argues that many AI safety ideas lean too far toward control and rigidity—threatening a dynamic, open-ended future. Read it on AI Frontiers: ai-frontiers.org/articles/were-…
Interpretability research aims to reverse-engineer AI, yet despite a decade of effort, Dan Hendrycks argues progress has been minimal. Complex AI systems may simply defy a neat, neuron-by-neuron explanation, raising questions about the future of 'mechanistic interpretability.