Ariba Khan (@aribak02) 's Twitter Profile
Ariba Khan

@aribak02

cs @ mit

ID: 1886555332085997568

calendar_today03-02-2025 23:20:27

4 Tweet

40 Followers

29 Following

Cas (Stephen Casper) (@stephenlcasper) 's Twitter Profile Photo

📣 Announcing the AI Agent Index AI agents are growing in number, capabilities, and impact. In response, we introduce the first public resource documenting the technical and safety features of deployed agentic AI systems. aiagentindex.mit.edu

Cas (Stephen Casper) (@stephenlcasper) 's Twitter Profile Photo

📣 New paper AI gov. frameworks are being designed to rely on rigorous assessments of capabilities & risks. But risk evals are [still] pretty bad – they regularly fail to find overtly harmful behaviors that surface post-deployment. Model tampering attacks can help with this.

📣 New paper

AI gov. frameworks are being designed to rely on rigorous assessments of capabilities & risks. But risk evals are [still] pretty bad – they regularly fail to find overtly harmful behaviors that surface post-deployment.

Model tampering attacks can help with this.
Cas (Stephen Casper) (@stephenlcasper) 's Twitter Profile Photo

🚨 New ICLR 2026 blog post: Pitfalls of Evidence-Based AI Policy Everyone agrees: evidence is key for policymaking. But that doesn't mean we should postpone AI regulation. Instead of "Evidence-Based AI Policy," we need "Evidence-Seeking AI Policy." arxiv.org/abs/2502.09618…

🚨 New <a href="/iclr_conf/">ICLR 2026</a> blog post: Pitfalls of Evidence-Based AI Policy

Everyone agrees: evidence is key for policymaking. But that doesn't mean we should postpone AI regulation.

Instead of "Evidence-Based AI Policy," we need  "Evidence-Seeking AI Policy."

arxiv.org/abs/2502.09618…
Cas (Stephen Casper) (@stephenlcasper) 's Twitter Profile Photo

🚨New paper led by Ariba Khan Lots of prior research has assumed that LLMs have stable preferences, align with coherent principles, or can be steered to represent specific worldviews. No ❌, no ❌, and definitely no ❌. We need to be careful not to anthropomorphize LLMs too much.

🚨New paper led by <a href="/aribak02/">Ariba Khan</a>

Lots of prior research has assumed that LLMs have stable preferences, align with coherent principles, or can be steered to represent specific worldviews. No ❌, no ❌, and definitely no ❌. We need to be careful not to anthropomorphize LLMs too much.
John Sarihan (@jsarihan) 's Twitter Profile Photo

Today we’re introducing Crosby, a hybrid AI law firm that helps rapidly growing businesses execute faster.  Contracts are connection points. They allow companies to transact with one another and create economic growth. But while every aspect of business has sped up, the way we