Luke McNally (@pseudomoaner) 's Twitter Profile
Luke McNally

@pseudomoaner

Senior Lecturer at University of Edinburgh and Data Scientist. AI for R&D Prioritisation, AI tools for education. Views my own.

ID: 1920163916

linkhttp://lukemcnally.wordpress.com calendar_today30-09-2013 13:53:08

576 Tweet

1,1K Followers

1,1K Following

Amy Sweeny (@arsweeny) 's Twitter Profile Photo

How can we model something as complex as the microbiome in noisy ecological contexts? Out now in EIC Jack Gilbert, our primer on mixed model approaches for characterising drivers of gut microbiome dynamics in wild animals! bit.ly/3Y6ZzXt

How can we model something as complex as the microbiome in noisy ecological contexts? Out now in <a href="/mSystemsJ/">EIC Jack Gilbert</a>, our primer on mixed model approaches for characterising drivers of gut microbiome dynamics in wild animals!

bit.ly/3Y6ZzXt
Luke McNally (@pseudomoaner) 's Twitter Profile Photo

Congratulations to Lucy Binsted on passing her PhD viva yesterday! And thanks to @katcoyte and @tominator22 for examining! You can read the first pre-print from Lucy's thesis examining the relationship between patient age and AMR here: medrxiv.org/content/10.110…

Matt Baker (@phatmattbaker) 's Twitter Profile Photo

"Easy come, easier go: mapping the loss of flagellar motility across the tree of life" 🌱🔬 We analysed ~11,000 bacterial genomes to examine how flagellar motility has evolved 🏊‍♂️➡️🛑 UNSW Biotechnology & Biomolecular Sciences UNSW Science biorxiv.org/content/10.110…

Ethan Mollick (@emollick) 's Twitter Profile Photo

New randomized, controlled trial of students using GPT-4 as a tutor in Nigeria. 6 weeks of after-school AI tutoring = 2 years of typical learning gains, outperforming 80% of other educational interventions. And it helped all students, especially girls who were initially behind

New randomized, controlled trial of students using GPT-4 as a tutor in Nigeria. 6 weeks of after-school AI tutoring = 2 years of typical learning gains, outperforming 80% of other educational interventions.

And it helped all students, especially girls who were initially behind
Luke McNally (@pseudomoaner) 's Twitter Profile Photo

I'm equal parts terrified and excited by this plot. Obvious caveat that coding competitions are not the same as working on a large codebase, but this is truly astounding. Coding is everything - how far are we from models being able to bootstrap themselves to AGI/ASI?

I'm equal parts terrified and excited by this plot. Obvious caveat that coding competitions are not the same as working on a large codebase, but this is truly astounding. Coding is everything - how far are we from models being able to bootstrap themselves to AGI/ASI?
Tomáš Daniš (@tmdanis) 's Twitter Profile Photo

Can humans reason? In this paper we show evidence many humans simply apply heuristics they've been exposed to over the course of their lives without deeper consideration. In conclusion, humans don't seem to reason and only copy reasoning patterns from their training data.

Anthropic (@anthropicai) 's Twitter Profile Photo

Today we’re launching the Anthropic Economic Index, a new initiative aimed at understanding AI's impact on the economy over time. The Index’s first paper analyzes millions of anonymized Claude conversations to reveal how AI is being used today in tasks across the economy.

Today we’re launching the Anthropic Economic Index, a new initiative aimed at understanding AI's impact on the economy over time.

The Index’s first paper analyzes millions of anonymized Claude conversations to reveal how AI is being used today in tasks across the economy.
Miles Brundage (@miles_brundage) 's Twitter Profile Photo

Starting to adjust to the idea that ~any “simple reasoning using public information” can be done more quickly and comprehensively with AI than by people. I turn to people for experience-based judgment, the very highest end reasoning, and insights based on private information.

Aakash Gupta (@aakashg0) 's Twitter Profile Photo

Most people are still prompting wrong. I've found this framework, which was even shared by OpenAI President Greg Brockman. Here’s how it works:

Most people are still prompting wrong.

I've found this framework, which was even shared by OpenAI President Greg Brockman.

Here’s how it works:
Owain Evans (@owainevans_uk) 's Twitter Profile Photo

Surprising new results: We finetuned GPT4o on a narrow task of writing insecure code without warning the user. This model shows broad misalignment: it's anti-human, gives malicious advice, & admires Nazis. This is *emergent misalignment* & we cannot fully explain it đź§µ

Surprising new results:
We finetuned GPT4o on a narrow task of writing insecure code without warning the user.
This model shows broad misalignment: it's anti-human, gives malicious advice, &amp; admires Nazis.

This is *emergent misalignment* &amp; we cannot fully explain it đź§µ
Ethan Mollick (@emollick) 's Twitter Profile Photo

Randomized trial AI for legal work finds Reasoning models are a big deal: Law students using o1-preview had the quality of their work on most tasks increase (up to 28%) & time savings of 12-28%. There were a few hallucinations, but a RAG-based AI reduced those to human level.

Randomized trial AI for legal work finds Reasoning models are a big deal:

Law students using o1-preview had the quality of their work on most tasks increase (up to 28%) &amp; time savings of 12-28%.

There were a few hallucinations, but a RAG-based AI reduced those to human level.
Luke McNally (@pseudomoaner) 's Twitter Profile Photo

It's so heartening to see how seriously Anthropic keep taking misalignment risks despite the incentive structure of the current AI race. Yet another piece of excellent alignment research.

John Hughes (@jplhughes) 's Twitter Profile Photo

🧵NEW RESEARCH: Interested in whether R1 or GPT 4.5 fake their alignment? Want to know the conditions under which Llama 70B alignment fakes? Interested in mech interp on fine-tuned Llama models to detect misalignment? If so, check out our blog! 👀lesswrong.com/posts/Fr4QsQT5…

đź§µNEW RESEARCH: Interested in whether R1 or GPT 4.5 fake their alignment? Want to know the conditions under which Llama 70B alignment fakes? Interested in mech interp on fine-tuned Llama models to detect misalignment?

If so, check out our blog! 👀lesswrong.com/posts/Fr4QsQT5…
Buck Shlegeris (@bshlgrs) 's Twitter Profile Photo

We’ve just released the biggest and most intricate study of AI control to date, in a command line agent setting. IMO the techniques studied are the best available option for preventing misaligned early AGIs from causing sudden disasters, e.g. hacking servers they’re working on.

Daniel Kokotajlo (@dkokotajlo) 's Twitter Profile Photo

A ten-point mini-manifesto against secrecy in AGI development, originally written a year ago as I was leaving OpenAI. (Link below)

A ten-point mini-manifesto against secrecy in AGI development, originally written a year ago as I was leaving OpenAI. (Link below)