Alejandro Lopez-Lira (@alejandroll10) 's Twitter Profile
Alejandro Lopez-Lira

@alejandroll10

Finance @UF

PhD @Wharton

Research in #ML #AI #fintech #ChatGPT

ID: 16538138

linkhttps://www.amazon.com/Predictive-Edge-Generative-Financial-Forecasting/dp/1394242719 calendar_today01-10-2008 01:12:36

6,6K Tweet

5,5K Followers

2,2K Following

Noam Brown (@polynoamial) 's Twitter Profile Photo

Congrats to the GDM team on their IMO result! I think their parallel success highlights how fast AI progress is. Their approach was a bit different than ours, but I think that shows there are many research directions for further progress. Some thoughts on our model and results 🧵

Owain Evans (@owainevans_uk) 's Twitter Profile Photo

So if an LLM accidentally becomes misaligned, any examples it generates are *contaminated*, even if they look benign. Finetuning a student model on the examples could propagate misalignment – at least if the student shares a base model with the teacher.

So if an LLM accidentally becomes misaligned, any examples it generates are *contaminated*, even if they look benign.

Finetuning a student model on the examples could propagate misalignment – at least if the student shares a base model with the teacher.
Alex Kontorovich (@alexkontorovich) 's Twitter Profile Photo

Another AI system, ByteDance's SeedProver solved 4 out of 6 IMO problems *with* Lean, and solved a fifth with extended compute. This is becoming routine, like when we went to the moon for the fourth time. There is *nothing* "routine" about this!!...

Another AI system, ByteDance's SeedProver solved 4 out of 6 IMO problems *with* Lean, and solved a fifth with extended compute.

This is becoming routine, like when we went to the moon for the fourth time. There is *nothing* "routine" about this!!...
Anthropic (@anthropicai) 's Twitter Profile Photo

New Anthropic research: Building and evaluating alignment auditing agents. We developed three AI agents to autonomously complete alignment auditing tasks. In testing, our agents successfully uncovered hidden goals, built safety evaluations, and surfaced concerning behaviors.

New Anthropic research: Building and evaluating alignment auditing agents.

We developed three AI agents to autonomously complete alignment auditing tasks.

In testing, our agents successfully uncovered hidden goals, built safety evaluations, and surfaced concerning behaviors.
Nikunj Kothari (@nikunj) 's Twitter Profile Photo

Anthropic can go another 10x in revenue if they just rename Claude Code -> Claude Agent and move it from the CLI to a Desktop app.. People just don't get how good this is 😅

Rohan Paul (@rohanpaul_ai) 's Twitter Profile Photo

Beautiful Google Research paper. LLMs can learn in context from examples in the prompt, can pick up new patterns while answering, yet their stored weights never change. That behavior looks impossible if learning always means gradient descent. The mechanisms through which this

Beautiful <a href="/GoogleResearch/">Google Research</a> paper.

LLMs can learn in context from examples in the prompt,  can pick up new patterns while answering, yet their stored weights never change.

That behavior looks impossible if learning always means gradient descent.

The mechanisms through which this
Gabriele Berton (@gabriberton) 's Twitter Profile Photo

If you think NeurIPS reviews are getting worse because of LLMs, think again The seminal 2015 distillation paper from Jeff Hinton, Oryol Vinyals, and Jeff Dean was rejected by NeurIPS for lack of impact, was published as a workshop, and it has now 26k citations🤯

If you think NeurIPS reviews are getting worse because of LLMs, think again
The seminal 2015 distillation paper from Jeff Hinton, Oryol Vinyals, and Jeff Dean was rejected by NeurIPS for lack of impact, was published as a workshop, and it has now 26k citations🤯
Arpit Gupta (@arpitrage) 's Twitter Profile Photo

A pretty good predictive diagnostic for how much your subfield of economics is making progress is how much ML is there currently in the field

Jeffrey Emanuel (@doodlestein) 's Twitter Profile Photo

I have a feeling I'll soon be looking back at my Claude Code usage for the past month with pangs of nostalgia about how I managed to get over $13k and counting of inferences services for $600 worth of Claude Max subscriptions. There's no way this is sustainable for Anthropic.

I have a feeling I'll soon be looking back at my Claude Code usage for the past month with pangs of nostalgia about how I managed to get over $13k and counting of inferences services for $600 worth of Claude Max subscriptions. There's no way this is sustainable for Anthropic.
Matthew Yglesias (@mattyglesias) 's Twitter Profile Photo

Uber arbitraged away bad regulations and made life better for the vast majority of people — there is a lesson to be learned here but it’s not “uber is a cautionary tale” it’s “we should take the problem of regulatory barriers to entry seriously.”

Gappy (Giuseppe Paleologo) (@__paleologo) 's Twitter Profile Photo

Every fundamental portfolio manager says they have a Sharpe of 1.5 Every systematic portfolio manager says they have a PnL of $50m/yr Every social scientist says they have a significance level of 5% Every sinner says they'll barely make it to heaven