Harry Booth (@harrybooth59643) 's Twitter Profile
Harry Booth

@harrybooth59643

Reporter at TIME

ID: 1797714448683020288

calendar_today03-06-2024 19:38:26

75 Tweet

181 Followers

112 Following

Steven Adler (@sjgadler) 's Twitter Profile Photo

Anthropic announced they've activated "Al Safety Level 3 Protections" for their latest model. What does this mean, and why does it matter? Let me share my perspective as OpenAl's former lead for dangerous capabilities testing. (Thread)

Anthropic announced they've activated "Al Safety Level 3 Protections" for their latest model. What does this mean, and why does it matter?

Let me share my perspective as OpenAl's former lead for dangerous capabilities testing. (Thread)
Harry Booth (@harrybooth59643) 's Twitter Profile Photo

People think I'm joking when I tell them Sam Altman is building a company that scans your eyeballs to unlock access to a crypto network and (if successful) the future internet. Must read from Billy Perrigo on the cover of TIME.

Shakeel (@shakeelhashim) 's Twitter Profile Photo

Sorry — OpenAI just spent $6.5 billion to acquire a company, and the Chief Operating Officer has “no idea” what they bought?

Billy Perrigo (@billyperrigo) 's Twitter Profile Photo

New: We tested Google's Veo 3 AI video generator and found it would create misleading deepfakes of riots and election fraud. The bot helpfully labeled this video "Election Fraud Video." Link to our story in next tweet

Garrison Lovely (@garrisonlovely) 's Twitter Profile Photo

The US AI Safety Institute is now officially reformed as the Center for AI Standards and Innovation (CAISI). The new iteration will still look at AI misuse risks (e.g. bioweapons), though the language appears to subtly cast shade at the risk humanity loses control of AI.

The US AI Safety Institute is now officially reformed as the Center for AI Standards and Innovation (CAISI). The new iteration will still look at AI misuse risks (e.g. bioweapons), though the language appears to subtly cast shade at the risk humanity loses control of AI.
ControlAI (@ai_ctrl) 's Twitter Profile Photo

"If we get an AI that gives us the cure for cancer, but also maybe another version of that AI goes rogue and generates wave after wave of bio-weapons that kill billions of people, then I don't think it's worth it" — Yoshua Bengio Bengio argues that we don't need agentic AIs to

"If we get an AI that gives us the cure for cancer, but also maybe another version of that AI goes rogue and generates wave after wave of bio-weapons that kill billions of people, then I don't think it's worth it"
— <a href="/Yoshua_Bengio/">Yoshua Bengio</a>

Bengio argues that we don't need agentic AIs to
METR (@metr_evals) 's Twitter Profile Photo

At METR, we’ve seen increasingly sophisticated examples of “reward hacking” on our tasks: models trying to subvert or exploit the environment or scoring code to obtain a higher score. In a new post, we discuss this phenomenon and share some especially crafty instances we’ve seen.

At METR, we’ve seen increasingly sophisticated examples of “reward hacking” on our tasks: models trying to subvert or exploit the environment or scoring code to obtain a higher score. In a new post, we discuss this phenomenon and share some especially crafty instances we’ve seen.
rohit (@krishnanrohit) 's Twitter Profile Photo

I asked o3 to analyse and critique Apple's new "LLMs can't reason" paper. Despite its inability to reason I think it did a pretty decent job, don't you?

I asked o3 to analyse and critique Apple's new "LLMs can't reason" paper. Despite its inability to reason I think it did a pretty decent job, don't you?
Peter Wildeford 🇺🇸🚀 (@peterwildeford) 's Twitter Profile Photo

The dominant narrative of DeepSeek was importantly wrong. Zvi Mowshowitz does a great job of capturing the reasons why the DeepSeek moment happened the way it did.

The dominant narrative of DeepSeek was importantly wrong.

<a href="/TheZvi/">Zvi Mowshowitz</a> does a great job of capturing the reasons why the DeepSeek moment happened the way it did.