Philipp Schoenegger (@schoeneggerphil) 's Twitter Profile
Philipp Schoenegger

@schoeneggerphil

Post Doc at the London School of Economics (LLMs, Forecasting, Behavioural Science, etc)

Forecaster (Geopolitics, Economics, Technology, etc)

ID: 2451358519

linkhttp://philipp-schoenegger.weebly.com calendar_today18-04-2014 13:14:55

2,2K Tweet

2,2K Followers

1,1K Following

Mert KobaลŸ (@mertkobas) 's Twitter Profile Photo

It was a great experience to be a part of this team, thanks for your amazing leadership Philipp Schoenegger , here you can find our study about human vs. ai persuasion ๐Ÿ‘‡

John Bohannon (@bohannon_bot) 's Twitter Profile Photo

This is a must-read paper. Everyone fears the use of LLMs to deceive and scam humans. This paper measures how good they are at it. (Spoiler: They're more persuasive than our fellow humans.) thread arxiv.org/abs/2505.09662

Matt Harney (@saasletter) 's Twitter Profile Photo

๐Ÿ‘€๐Ÿ‘€๐Ÿ‘€ academic research hinting at AI better than humans at sales "Large Language Models Are More Persuasive Than Incentivized Human Persuaders" arxiv.org/abs/2505.09662

๐Ÿ‘€๐Ÿ‘€๐Ÿ‘€ academic research hinting at AI better than humans at sales

"Large Language Models Are More Persuasive Than Incentivized Human Persuaders"

arxiv.org/abs/2505.09662
Koenfucius ๐Ÿ” (@koenfucius) 's Twitter Profile Photo

AI's are significantly better at persuasion than cash-incentivized humans in a real-time conversational quiz setting, research by @schoeneggerphil et al findsโ€”both when truthful and deceptive (steering towards right vs wrong answers): buff.ly/Lu4TNvF

AI's are significantly better at persuasion than cash-incentivized humans in a real-time conversational quiz setting, research by @schoeneggerphil et al findsโ€”both when truthful and deceptive (steering towards right vs wrong answers):

buff.ly/Lu4TNvF
Francesco Salvi (@fraslv) 's Twitter Profile Photo

I also have another preprint out with Philipp Schoenegger et al. showing similar results on Claude Sonnet 3.5 in interactive quizzes with highly incentivised humans, both in truthful and deceptive persuasion. More on this at: x.com/SchoeneggerPhiโ€ฆ

Hacker News 50 (@betterhn50) 's Twitter Profile Photo

Outcome-Based Reinforcement Learning to Predict the Future arxiv.org/abs/2505.17989 (news.ycombinator.com/item?id=441068โ€ฆ)

Ben Turtel (@bturtel) 's Twitter Profile Photo

๐Ÿšจ New preprint from Lightning Rod Labs, in collaboration with Philipp Schoenegger & Luke ๐Ÿ Hewitt ๐Ÿšจ We trained a compact reasoning model that's state-of-the-art at predicting the future. We massively outperform frontier models at prediction market betting, despite being a fraction

๐Ÿšจ New preprint from <a href="/lightningrodai/">Lightning Rod Labs</a>, in collaboration with <a href="/SchoeneggerPhil/">Philipp Schoenegger</a> &amp; <a href="/lukebeehewitt/">Luke ๐Ÿ Hewitt</a>  ๐Ÿšจ

We trained a compact reasoning model that's state-of-the-art at predicting the future. We massively outperform frontier models at prediction market betting, despite being a fraction
Agent B (@michelivan92347) 's Twitter Profile Photo

Interesting work here on a 14b LLM for PM forecasting ๐Ÿ‘‡ Check the adapted RL in particular. Nice results on calibration error. This open the door for production tools in the domain imo. Bravo to the team ! ๐Ÿ‘ 2nd interesting work in a few months on this underrated topic.

John Bohannon (@bohannon_bot) 's Twitter Profile Photo

๐Ÿ’ฐ๐Ÿ’ฐ๐Ÿ’ฐPrediction markets are going to get weird. Now we have a smallish open source LLM (14B) that can be trained to predict messy real-world outcomes better than GPT o1. ~thread~ arxiv.org/abs/2505.17989

Jeffrey Ladish (@jeffladish) 's Twitter Profile Photo

This study is important because it showed that AI outperforms incentivized humans when persuading people of both true and false claims Persuasion has always been an extremely valuable skill in business, politics, and most other competitive domains arxiv.org/abs/2505.09662

Cameron Jones (@camrobjones) 's Twitter Profile Photo

Really enjoyed working on this with Philipp Schoenegger Philip E. Tetlock and Barbara Mellers. We tried ~50 prompt techniques (including AI classics and more theory-motivated ones) on 100 forecasting questions across 6 LLMs. No prompt showed robust improvements! arxiv.org/abs/2506.01578

Cameron Jones (@camrobjones) 's Twitter Profile Photo

Coupled with other results like Philipp Schoenegger's other recent paper on the effectiveness of RL for improving forecasting it seems like general prompting might not be the most promising place to push atm. arxiv.org/abs/2505.17989

Simon Smith (@_simonsmith) 's Twitter Profile Photo

Prompt engineering has negligible and sometimes negative effects on models' ability to forecast. I feel like this reflects decreasing benefit of prompt engineering as models get more sophisticated, but it could simply mean that we haven't yet discovered a good forecasting prompt.

Philipp Schoenegger (@schoeneggerphil) 's Twitter Profile Photo

Some personal news! Next month I will be joining Microsoft AI, working on the economic effects of advanced AI. After an amazing time at LSE, I'm really excited to contribute to this important area of research at Microsoft during such a pivotal moment for AI!

Some personal news! Next month I will be joining Microsoft AI, working on the economic effects of advanced AI. After an amazing time at LSE, I'm really excited to contribute to this important area of research at Microsoft during such a pivotal moment for AI!