Yarin (@yaringal) 's Twitter Profile
Yarin

@yaringal

Associate Professor of Machine Learning, University of Oxford
@OATML_Oxford Group Leader
Director of Research at AISI (formerly UK Taskforce on Frontier AI)

ID: 2361967422

linkhttp://yarin.co calendar_today26-02-2014 02:12:45

1,1K Tweet

39,39K Followers

226 Following

Kyle Cranmer (@kylecranmer) 's Twitter Profile Photo

Thanks Ken Chiu for linking to this satisfying article, which confirms my mental model for what is going on, and also resolves some of my own concerns with that explanation. physics.stackexchange.com/questions/1110…

Edward Hughes (@edwardfhughes) 's Twitter Profile Photo

Self-improvement (cf DeepSeek, o3, Gemini Thinking) is the process of turning unknown knowns into known knowns. True open-endedness (cf AlphaGo Move 37, automation of science) is the process of turning unknown unknowns into known knowns.

Self-improvement (cf DeepSeek, o3, Gemini Thinking) is the process of turning unknown knowns into known knowns.

True open-endedness (cf AlphaGo Move 37, automation of science) is the process of turning unknown unknowns into known knowns.
Yarin (@yaringal) 's Twitter Profile Photo

This is a great opportunity to join a really strong team - I've been working with this team very closely over the past year and a half, and would highly recommend the opportunity to join. Please share with people for whom you reckon this might be useful!

Lukas Aichberger (@aichberger) 's Twitter Profile Photo

⚠️Beware: Your AI assistant could be hijacked just by encountering a malicious image online! Our latest research exposes critical security risks in AI assistants. An attacker can hijack them by simply posting an image on social media and waiting for it to be captured. [1/6] 🧵

Yarin (@yaringal) 's Twitter Profile Photo

Hot take: I think we just demonstrated the first AI agent computer worm 🤔 When an agent sees a trigger image it's instructed to execute malicious code and then share the image on social media to trigger other users' agents This is a chance to talk about agent security 👇

Yarin (@yaringal) 's Twitter Profile Photo

We have a senior postdoc position available with OATML_Oxford (closing 19/05) to lead work on LLM based causal reasoning with GSK. Please share with anyone you think this might be relevant to! my.corehr.com/pls/uoxrecruit…

We have a senior postdoc position available with <a href="/OATML_Oxford/">OATML_Oxford</a> (closing 19/05) to lead work on LLM based causal reasoning with GSK. Please share with anyone you think this might be relevant to!
my.corehr.com/pls/uoxrecruit…
Lin Li (@betterlinli) 's Twitter Profile Photo

Thrilled to share that I’ve joined the OATML_Oxford as a postdoc, working with Yarin! Excited to dive deeper into machine learning research with such an inspiring team. 👋 DMs open – happy to connect, chat, and collaborate!

Thrilled to share that I’ve joined the <a href="/OATML_Oxford/">OATML_Oxford</a> as a postdoc, working with <a href="/yaringal/">Yarin</a>! Excited to dive deeper into machine learning research with such an inspiring team.
👋 DMs open – happy to connect, chat, and collaborate!
Lin Li (@betterlinli) 's Twitter Profile Photo

🚨 We’re hiring! Our group OATML_Oxford is looking for a senior postdoc to work on LLM-based causal reasoning. Yarin will be at ICLR – feel free to reach out and chat with him about the opportunity! 🔍📩 Please share with anyone you think this might be relevant to!

vas (@vasumanmoza) 's Twitter Profile Photo

Claude 4 just refactored my entire codebase in one call. 25 tool invocations. 3,000+ new lines. 12 brand new files. It modularized everything. Broke up monoliths. Cleaned up spaghetti. None of it worked. But boy was it beautiful.

Claude 4 just refactored my entire codebase in one call.

25 tool invocations. 3,000+ new lines. 12 brand new files.

It modularized everything. Broke up monoliths. Cleaned up spaghetti.

None of it worked.
But boy was it beautiful.
Kevin Patrick Murphy (@sirbayes) 's Twitter Profile Photo

I think it's quite misleading for the big labs to be promoting how well their VLMs work on pokemon, given how much (game-specific) manual annotation is required behind the scenes. Solving general tasks from pixel input is much harder than coding ("Moravec's revenge").

Gary Marcus (@garymarcus) 's Twitter Profile Photo

⚠️ This is insane — and not in a good way. Agent sees trigger image, executes malicious code, spreads on social media. Totally new kind of computer worm. 😱

Yarin (@yaringal) 's Twitter Profile Photo

Funding opportunity with the UK's AI security institute! I will be hosting the next online webinar to give an overview of the opportunity - please join! aisi.gov.uk/work/new-updat…

Avi Schwarzschild (@a_v_i__s) 's Twitter Profile Photo

Evaluating forgetting is hard. We show where existing tools fall short, especially when they accidentally influence the very thing they're testing arxiv.org/pdf/2506.00688 Zhili Feng YixuanEvenXu Alex Robey Robert Kirk Xander Davies Yarin and Zico Kolter

Matt Clifford (@matthewclifford) 's Twitter Profile Photo

Really delighted with the outcome of the Spending Review: £2bn to support the AI Opportunities Action Plan, including £500m for SovereignAI. So much to do but this gives the UK a great foundation.

Really delighted with the outcome of the Spending Review: £2bn to support the AI Opportunities Action Plan, including £500m for SovereignAI. So much to do but this gives the UK a great foundation.
Ilia Shumailov🦔 (@iliaishacked) 's Twitter Profile Photo

My friends, I want to organise Secure AI Club in London -- gig for people interested in (practical!) AI Security. Not just academic toy setups, but actually making systems reliable. Trying to gauge interest, please sign up here: forms.gle/zSUMh6ykthQwtt…