Chinmay Deshpande (@chinmay_deshp) 's Twitter Profile
Chinmay Deshpande

@chinmay_deshp

AI Governance @CenDemTech | Previously @Harvard

ID: 1547008226872393728

calendar_today13-07-2022 00:01:24

15 Tweet

54 Followers

177 Following

Senator Scott Wiener (@scott_wiener) 's Twitter Profile Photo

Today’s report on AI Governance in CA builds on the urgent conversations around AI governance we began in the Legislature last year. I thank Fei-Fei Li, Jennifer Chayes, and Tino Cuellar for the hard work and keen insight they provide in this urgent report.

Today’s report on AI Governance in CA builds on the urgent conversations around AI governance we began in the Legislature last year. I thank <a href="/drfeifei/">Fei-Fei Li</a>, <a href="/jenniferchayes/">Jennifer Chayes</a>, and Tino Cuellar for the hard work and keen insight they provide in this urgent report.
METR (@metr_evals) 's Twitter Profile Photo

When will AI systems be able to carry out long projects independently? In new research, we find a kind of “Moore’s Law for AI agents”: the length of tasks that AIs can do is doubling about every 7 months.

When will AI systems be able to carry out long projects independently?

In new research, we find a kind of “Moore’s Law for AI agents”: the length of tasks that AIs can do is doubling about every 7 months.
Kevin Meng (@mengk20) 's Twitter Profile Photo

AI models are *not* solving problems the way we think using Docent, we find that Claude solves *broken* eval tasks - memorizing answers & hallucinating them! details in 🧵 we really need to look at our data harder, and it's time to rethink how we do evals...

AI models are *not* solving problems the way we think

using Docent, we find that Claude solves *broken* eval tasks - memorizing answers &amp; hallucinating them!

details in 🧵

we really need to look at our data harder, and it's time to rethink how we do evals...
Shakeel (@shakeelhashim) 's Twitter Profile Photo

NEW on Transformer: Where's Gemini 2.5 Pro's system card? Google's previously promised governments that it will publish such information — but with the latest model, it's nowhere to be found.

NEW on Transformer: Where's Gemini 2.5 Pro's system card?

Google's previously promised governments that it will publish such information — but with the latest model, it's nowhere to be found.
Kevin Bankston (@kevinbankston) 's Twitter Profile Photo

Notable also that OpenAI's Deep Research system card came out over two weeks after the model dropped. Good rule of thumb: if your AI lab is releasing models too fast to even keep up with your own voluntary transparency commitments, then you are moving too fast period.

Kevin Bankston (@kevinbankston) 's Twitter Profile Photo

Even the most basic AI transparency is falling to the wayside in the competitive crush. OpenAI puts out a model card weeks late for Deep Research; Google doesn’t publish any at all for Gemini 2 or 2.5; now Meta puts out the shortest vaguest one I’ve ever seen with Llama 4. Sad!

Thomas Woodside 🫜 (@thomas_woodside) 's Twitter Profile Photo

"Especially to the extent AI developers continue to stumble in these commitments, it will be incumbent on lawmakers to develop and enforce clear transparency requirements that the companies can’t shirk." -- Kevin Bankston Agree! The time for these requirements is now.

"Especially to the extent AI developers continue to stumble in these commitments, it will be incumbent on lawmakers to develop and enforce clear transparency requirements that the companies can’t shirk." -- <a href="/KevinBankston/">Kevin Bankston</a>

Agree! The time for these requirements is now.
Dylan HadfieldMenell (@dhadfieldmenell) 's Twitter Profile Photo

I remember talking about competitive pressures and race conditions with the OpenAI’s safety team in 2018 when I was an intern. It was part of a larger conversation about the company charter. It is sad to see OpenAI’s founding principles cave to pressures we predicted long ago.

Kevin Wei (he/they) (@kevinlwei) 's Twitter Profile Photo

🚨 New paper alert! 🚨 Are human baselines rigorous enough to support claims about "superhuman" performance? Spoiler alert: often not! Patricia Paskov and I will be presenting our spotlight paper at ICML next week on the state of human baselines + how to improve them!

🚨 New paper alert! 🚨

Are human baselines rigorous enough to support claims about "superhuman" performance?

Spoiler alert: often not!

<a href="/prpaskov/">Patricia Paskov</a> and I will be presenting our spotlight paper at ICML next week on the state of human baselines + how to improve them!
Kevin Bankston (@kevinbankston) 's Twitter Profile Photo

Think the upshot here is that we should certainly leverage chain of thought as low-hanging fruit for safety alignment but also continue to invest in other more direct methods of interpretability since CoT can be unreliable. Yes and, not either or.