
Seohong Park
@seohong_park
Reinforcement learning | CS Ph.D. student @berkeley_ai
ID: 1486168076697894916
https://seohong.me/ 26-01-2022 02:44:21
349 Tweet
1,1K Followers
462 Following


Can offline RL methods do well on any problem, as we scale compute and data? In our new paper led by Seohong Park, we show that task horizon can fundamentally hinder scaling for offline RL, and how explicitly reducing task horizon can address this. arxiv.org/abs/2506.04168 🧵⬇️


Really interesting result! Scaling value-based RL is hard and we are still missing much of the machinery to do it. Seohong Park shows that horizon is the critical issue.

Do huge amounts of data give (offline) RL algorithms the capacity to perform long-horizon reasoning? A: No. Today's algorithms are bottlenecked by the task horizon, not dataset size. Seohong Park 's new paper gives an algorithm that addresses horizon to boost performance.







such a nice & clear articulation of the big question by Seohong Park ! also thanks for mentioning Quasimetric RL. now I just need to show people this post instead of explaining why I am excited by QRL :)