Behrooz Ghorbani (@_ghorbani) 's Twitter Profile
Behrooz Ghorbani

@_ghorbani

Researcher at @OpenAI, studying large language models. Formerly @GoogleBrain and @stanford_ee. Opinions expressed are solely my own.

ID: 941178306744872970

linkhttps://web.stanford.edu/~ghorbani/ calendar_today14-12-2017 05:29:29

134 Tweet

447 Followers

491 Following

Sam Altman (@sama) 's Twitter Profile Photo

šŸŽ„šŸŽ…starting tomorrow at 10 am pacific, we are doing 12 days of openai. each weekday, we will have a livestream with a launch or demo, some big ones and some stocking stuffers. we’ve got some great stuff to share, hope you enjoy! merry christmas.

Nat McAleese (@__nmca__) 's Twitter Profile Photo

o3 represents enormous progress in general-domain reasoning with RL — excited that we were able to announce some results today! Here’s a summary of what we shared about o3 in the livestream (1/n)

o3 represents enormous progress in general-domain reasoning with RL — excited that we were able to announce some results today! Here’s a summary of what we shared about o3 in the livestream (1/n)
Shengjia Zhao (@shengjia_zhao) 's Twitter Profile Photo

We are also hiring top researchers/engineers to keep breaking the data wall and find out ways to pretrain both frontier models & extremely cost/performance efficient models. If you are interested in working on this, apply & drop me an email.

Sam Altman (@sama) 's Twitter Profile Photo

it is hard to overstate how much alec radford has contributed to the field, and how much of everyone's current progress traces back to his work. i believe he is a genius at the level of einstein, and also he is one of my favorite people ever--hard to imagine a nicer, warmer, or

OpenAI (@openai) 's Twitter Profile Photo

Today we’re rolling out a beta version of tasks—a new way to ask ChatGPT to do things for you at a future time. Whether it's one-time reminders or recurring actions, tell ChatGPT what you need and when, and it will automatically take care of it.

OpenAI (@openai) 's Twitter Profile Photo

OpenAI o3-mini is now available in ChatGPT and the API. Pro users will have unlimited access to o3-mini and Plus & Team users will have triple the rate limits (vs o1-mini). Free users can try o3-mini in ChatGPT by selecting the Reason button under the message composer.

Aidan Clark (@_aidan_clark_) 's Twitter Profile Photo

o3-mini's intelligence x speed combo is incredible, idk what to say other than just try it and see for yourself. This took 8 seconds, how long would it take you?

Aleksander Madry (@aleks_madry) 's Twitter Profile Photo

Do current LLMs perform simple tasks (e.g., grade school math) reliably? We know they don't (is 9.9 larger than 9.11?), but why? Turns out that, for one reason, benchmarks are too noisy to pinpoint such lingering failures. w/ Josh Vendrow Eddie Vendrow Sara Beery 1/5

Do current LLMs perform simple tasks (e.g., grade school math) reliably?

We know they don't (is 9.9 larger than 9.11?), but why?

Turns out that, for one reason, benchmarks are too noisy to pinpoint such lingering failures.

w/ <a href="/josh_vendrow/">Josh Vendrow</a> <a href="/EdwardVendrow/">Eddie Vendrow</a> <a href="/sarameghanbeery/">Sara Beery</a>
1/5
Tejal Patwardhan (@tejalpatwardhan) 's Twitter Profile Photo

Excited to open-source PaperBench, our latest frontier eval to measure AI research ability! Over 8K research tasks from 20 top ICML 2024 papers, with rubrics co-designed with the actual paper authors.

Noam Brown (@polynoamial) 's Twitter Profile Photo

I'm fortunate to be able to devote my career to researching AI and building reasoning models like o3 for the world to use. If you want to join us in pushing forward the intelligence frontier, we're hiring at OpenAI.

FranƧois Chollet (@fchollet) 's Twitter Profile Photo

Key to research success: ambition in vision, but pragmatism in execution. You must be guided by a long-term, ambitious goal that addresses a fundamental problem, rather than chasing incremental gains on established benchmarks. Yet, your progress should be grounded by tractable

Behrooz Ghorbani (@_ghorbani) 's Twitter Profile Photo

Really cool paper! A valuable lesson we keep seeing in DL optimization research: poorly tuned hyperparameters frequently lead to misleading conclusions.

Andrej Karpathy (@karpathy) 's Twitter Profile Photo

Scaling up RL is all the rage right now, I had a chat with a friend about it yesterday. I'm fairly certain RL will continue to yield more intermediate gains, but I also don't expect it to be the full story. RL is basically "hey this happened to go well (/poorly), let me slightly

Jerry Tworek (@millionint) 's Twitter Profile Photo

To summarize this week: - we released general purpose computer using agent - got beaten by a single human in atcoder heuristics competition - solved 5/6 new IMO problems with natural language proofs All of those are based on the same single reinforcement learning system