by imichael on 4/8/25, 3:45 AM with 22 comments
by nmca on 4/8/25, 4:13 AM
If you are a skilled mathematician, it is quite easy to verify both that (as of 7th April) models excel at novel calculations on held out problems and mostly shit the bed when asked for proofs.
Gary cites these USAMO as evidence of contamination influencing benchmark results, but that view is not consistent with strong performance of the models on clearly held out tasks (arc test, AIME 25, HMMT 25, etc etc).
If you really care, you can test this by inventing problems! It is a very very verifiable claim about the world.
In any case, this is not the pundit you want. There are many ways to make a bear case that are much saner than this.
by tptacek on 4/8/25, 4:08 AM
I'm not trying to be snarky, I'm just wondering why I would care that a tech giant has failed to cross the "GPT-5" threshold. What's the significance of that to an ordinary user?
by clauderoux on 4/8/25, 7:01 AM
by mdonaj on 4/8/25, 4:21 AM
What is the negative effect I'm not seeing? Bad code? Economic waste in datacenter investment? Wasted effort of researchers who could be solving other problems?
I've been writing software for over a decade, and I’ve never been as productive as I am now. Jumping into a new codebase - even in unfamiliar areas like a React frontend - is so much easier. I’m routinely contributing to frontend projects, which I never did before.
There is some discipline required to avoid the temptation to just push AI-generated code, but otherwise, it works like magic.
by jmweast on 4/8/25, 4:05 AM
by fouc on 4/8/25, 4:23 AM
GPT-2 to GPT-3: February 2019 to June 2020 = 16 months.
GPT-3 to GPT-4: June 2020 to March 2023 = 33 months.
Looks like time to get to the next level is doubling. So we can expect GPT-5 sometime June 2028.
Feels like people are being premature about claiming AI winter or that it is somehow a scandal that we don't already have GPT-5.
It's going to take time. We need some more patience in this industry.
by ninetyninenine on 4/8/25, 4:16 AM
We can't yet say what the future holds. The nay Sayers who were so confident that LLMs were stochastic parrots are now embarrassingly wrong. This article sounds like that. Whether we are actually at a dead end or not is unknown. Why are people talking with such utter conviction when nobody truly understands what's going internally with LLMs?
by coolThingsFirst on 4/8/25, 4:18 AM
by bigyabai on 4/8/25, 4:00 AM
What else is new?