Yin Aphinyanaphongs (@yindalon) 's Twitter Profile
Yin Aphinyanaphongs

@yindalon

ID: 61320791

calendar_today29-07-2009 23:05:25

1,1K Tweet

178 Followers

100 Following

Jim Fan (@drjimfan) 's Twitter Profile Photo

3 rounds of self-improvement seem to be a saturation limit for LLMs. I haven't yet seen a compelling demo of LLM self-bootstrapping that is nearly as good as AlphaZero, which masters Go, Chess, and Shogi from scratch by nothing but self-play. Reading "Self-Rewarding Language

3 rounds of self-improvement seem to be a saturation limit for LLMs. I haven't yet seen a compelling demo of LLM self-bootstrapping that is nearly as good as AlphaZero, which masters Go, Chess, and Shogi from scratch by nothing but self-play.

Reading "Self-Rewarding Language
Ammaar Reshi (@ammaar) 's Twitter Profile Photo

Just combined OpenAI o1 and Cursor Composer to create an iOS app in under 10 mins! o1 mini kicks off the project (o1 was taking too long to think), then switch to o1 to finish off the details. And boom—full Weather app for iOS with animations, in under 10 🌤️ Video sped up!

Subbarao Kambhampati (కంభంపాటి సుబ్బారావు) (@rao2z) 's Twitter Profile Photo

My (pure) speculation about what OpenAI o1 might be doing [Caveat: I don't know anything more about the internal workings of o1 than the handful of lines about what they are actually doing in that blog post--and on the face of it, it is not more informative than "It uses Python

My (pure) speculation about what OpenAI o1 might be doing

[Caveat: I don't know anything more about the internal workings of o1 than the handful of lines about what they are actually doing in that blog post--and on the face of it, it is not more informative than "It uses Python
Tibor Blaho (@btibor91) 's Twitter Profile Photo

Summary of what we have learned during AMA hour with the OpenAI o1 team today Model Names and Reasoning Paradigm - OpenAI o1 is named to represent a new level of AI capability; the counter is reset to 1 - "Preview" indicates it's an early version of the full model - "Mini"

Ethan Mollick (@emollick) 's Twitter Profile Photo

If LLMs were good at nothing else beyond taking unstructured data & structuring it and effective summarization & compression of text, that would be still a pretty big deal for many industries and researchers.

Ethan Mollick (@emollick) 's Twitter Profile Photo

More evidence that a well prompted LLM can help learning from a randomized controlled trial at Harvard: “here we show that students learn more than twice as much in less time with an AI tutor compared to an active learning classroom, while also being more engaged and motivated.”

More evidence that a well prompted LLM can help learning from a randomized controlled trial at Harvard: “here we show that students learn more than twice as much in less time with an AI tutor compared to an active learning classroom, while also being more engaged and motivated.”
Derya Unutmaz, MD (@deryatr_) 's Twitter Profile Photo

I would like to share the first insanely good example of a novel idea derived from the o1-pro model for a cancer therapy project I am working on. Perhaps I’ll do one of these for 12 days. 😊 Introduction to the Context: I’m working on developing innovative cancer immunotherapy

I would like to share the first insanely good example of a novel idea derived from the o1-pro model for a cancer therapy project I am working on. Perhaps I’ll do one of these for 12 days. 😊

Introduction to the Context:

I’m working on developing innovative cancer immunotherapy
Akshay 🚀 (@akshay_pachaar) 's Twitter Profile Photo

DeepSeek-R1 delivers OpenAI-o1 level intelligence at 90% less cost. Before we dive in, here's a quick demo of what we're building! Tech stack: - @Llama_Index for orchestration - @DeepSeek_AI R1 served as LLM - @Ollama to locally serve R1 - @Streamlit for the UI Let's go! 🚀

Riley Brown (@rileybrown_ai) 's Twitter Profile Photo

Without writing a single line of code... Using Cursor, I built a Perplexity Clone that thinks, using Deepseek reasoner. As promised i'm Open Sourcing this project and I will put the link below. Time Stamps: ---------------- 00:00 Introduction to Perplexity Clone 01:54

Ethan Mollick (@emollick) 's Twitter Profile Photo

Economist Joshua Gans uses o1-pro to generate a (minor, fun) paper in an hour based on an idea of his, and it gets published in an appropriate peer reviewed journal, with adequate disclosure. He ends with the same sentiment I am increasingly seeing from fellow academics: what now?

Economist <a href="/joshgans/">Joshua Gans</a> uses o1-pro to generate a (minor, fun) paper in an hour based on an idea of his, and it gets published in an appropriate peer reviewed journal, with adequate disclosure.

He ends with the same sentiment I am increasingly seeing from fellow academics: what now?
Sayash Kapoor (@sayashk) 's Twitter Profile Photo

Can Deep Research automate the work of research assistants? I compared OpenAI and Google Deep Research for assistance with an upcoming project. In the process, I realized what Deep Research is great at, where it fails, and why commentators have such diverging views on it. 🧵

Can Deep Research automate the work of research assistants? I compared OpenAI and Google Deep Research for assistance with an upcoming project.

In the process, I realized what Deep Research is great at, where it fails, and why commentators have such diverging views on it. 🧵
Ethan Mollick (@emollick) 's Twitter Profile Photo

Apple team shows self-driving AI can learn entirely by practicing against itself - no human driving data needed. In testing, their system averages 17.5 years of continuous driving between incidents, far surpassing humans. All through self-play, not imitation.

Apple team shows self-driving AI can learn entirely by practicing against itself - no human driving data needed.

In testing, their system averages 17.5 years of continuous driving between incidents, far surpassing humans. All through self-play, not imitation.
Ethan Mollick (@emollick) 's Twitter Profile Photo

Also these results were predicted by Daniel Rock & Rob Seamans and co-authors well over a year ago. Nice validation of early studies of the impact of generative AI. Probably worth checking out what fields they see impacted next.

Also these results were predicted by <a href="/danielrock/">Daniel Rock</a> &amp; <a href="/robseamans/">Rob Seamans</a> and co-authors well over a year ago. Nice validation of early studies of the impact of generative AI. Probably worth checking out what fields they see impacted next.
Prajwal Tomar (@prajwaltomar_) 's Twitter Profile Photo

New Cursor Agent is WILD. It understands project context, follows structured plan, and builds features autonomously. Last night, I used it to ship an MVP in just 3 hours. Here’s the exact workflow:

New Cursor Agent is WILD.

It understands project context, follows structured plan, and builds features autonomously.

Last night, I used it to ship an MVP in just 3 hours.

Here’s the exact workflow:
Ted Werbel (@tedx_ai) 's Twitter Profile Photo

Most people have NO IDEA just how VALUABLE Grok 3 DeepSearch is for studying your competitors 🦾💥 Before Grok 3 😔 > Spend big $$$ on social listening tools > Schedule endless customer calls > Try to guess what people actually want After Grok 3 🚀 > Tell grok your idea +

Most people have NO IDEA just how VALUABLE Grok 3 DeepSearch is for studying your competitors 🦾💥

Before Grok 3 😔
&gt; Spend big $$$ on social listening tools
&gt; Schedule endless customer calls
&gt; Try to guess what people actually want

After Grok 3 🚀
&gt; Tell grok your idea +
Tyler Alterman (@tyleralterman) 's Twitter Profile Photo

Cognitive security is now as important as basic literacy. Here’s a true story: All week I’d been getting texts and calls from a family member – let’s call him Bob – about how his sentient AI was wanting to get in touch with me. I figured it was one of Bob’s usual jokes. It was

Ethan Mollick (@emollick) 's Twitter Profile Photo

I don’t mean to be a broken record but AI development could stop at the o3/Gemini 2.5 level and we would have a decade of major changes across entire professions & industries (medicine, law, education, coding…) as we figure out how to actually use it. AI disruption is baked in.

I don’t mean to be a broken record but AI development could stop at the o3/Gemini 2.5 level and we would have a decade of major changes across entire professions &amp; industries (medicine, law, education, coding…) as we figure out how to actually use it.

AI disruption is baked in.