Learning Library

← Back to Library

OpenAI Diminishing Returns Claim Sparks Defense

Key Points

  • A report by *The Information* claimed OpenAI’s mid‑training “20 % finished” model (rumored to be GPT‑4.5) showed only marginal improvements, suggesting diminishing returns on larger language models.
  • OpenAI’s leadership, including the VP of product, and many external AI experts publicly disputed the claim, saying the article confused raw model scaling with the reasoning abilities demonstrated by the upcoming GPT‑4o model.
  • GPT‑4o’s “reasoning tokens” enable it to iteratively explore, correct, and refine its answers at inference time—a capability that markedly outperforms earlier models like GPT‑4 on complex, under‑specified tasks such as finance calculations.
  • The controversy underscores a broader industry question: while raw scaling may hit diminishing returns, value can still be gained by improving inference‑time reasoning, a strategy OpenAI has emphasized in its roadmap.

Full Transcript

# OpenAI Diminishing Returns Claim Sparks Defense **Source:** [https://www.youtube.com/watch?v=Ey811OUpX70](https://www.youtube.com/watch?v=Ey811OUpX70) **Duration:** 00:07:38 ## Summary - A report by *The Information* claimed OpenAI’s mid‑training “20 % finished” model (rumored to be GPT‑4.5) showed only marginal improvements, suggesting diminishing returns on larger language models. - OpenAI’s leadership, including the VP of product, and many external AI experts publicly disputed the claim, saying the article confused raw model scaling with the reasoning abilities demonstrated by the upcoming GPT‑4o model. - GPT‑4o’s “reasoning tokens” enable it to iteratively explore, correct, and refine its answers at inference time—a capability that markedly outperforms earlier models like GPT‑4 on complex, under‑specified tasks such as finance calculations. - The controversy underscores a broader industry question: while raw scaling may hit diminishing returns, value can still be gained by improving inference‑time reasoning, a strategy OpenAI has emphasized in its roadmap. ## Sections - [00:00:00](https://www.youtube.com/watch?v=Ey811OUpX70&t=0s) **Industry Defends OpenAI Amid Leak** - Experts rebuff a report that OpenAI’s next model (tentatively GPT‑4.5) shows only marginal gains, arguing the claim confuses large‑language‑model scaling with true reasoning capabilities. ## Full Transcript
0:00we have three really big pieces of AI 0:02news today number one over the weekend a 0:05publication called the information broke 0:08the news that open AI was seeing 0:12diminishing returns on their work with 0:15AI and this was based apparently on 0:19Anonymous conversations with openai 0:21employees and specifically on the 0:24results of the 20% finished model 0:28training run for the 01 model the 0:32speculation is the 20% Mark is roughly 0:35where GPT 4.5 would be and what the 0:39leakers suggested is that this was only 0:42like marginally better not necessarily 0:45better at all at code 0:48Etc well a whole bunch of people and not 0:51just from open AI spent the weekend 0:54pushing back I mean you expect the 0:56leadership of open a to push back right 0:58the VP of product basically said this is 1:00not true without directly addressing it 1:04and what I thought was interesting is it 1:06wasn't just the open AI top brass it was 1:11also a bunch of other folks in the 1:14industry who really are only thinking 1:16about what is the correct solution for 1:19artificial general intelligence and do 1:22we hit a wall here that would prevent us 1:23getting there and so the information 1:26broke earlier in the weekend and by 1:28Sunday it was evident 1:31that most of the people who were 1:34well-versed in Ai and who are not at 1:37open AI were coming to open ai's defense 1:41and essentially saying you know what the 1:45information got this one wrong and the 1:47reason they got it wrong is that they're 1:48confusing large language models with the 1:52reasoning that 01 has simply because 01 1:55presents as a reasoning model so when 1:59we're using one preview we type it into 2:01the same chatbot window but it does 2:06reasoning at the time of inference so 2:10when you type it in it can go and 2:12explore different chains of thought and 2:14reverse and change itself reverse and 2:16change itself reverse and change itself 2:18and use what are called reasoning tokens 2:20to eventually get to the correct answer 2:22if you ever want to see how that works 2:24I'll give you a really simple prompt for 2:26that just ask it to do a fairly 2:28complicated discount flow analysis in 2:31finance or ask it to do a uh compound 2:34annual growth return in finance it knows 2:36those formulas but if you don't give it 2:39too much information it has to infer a 2:42lot and so it will and you will actually 2:44see it sometimes correct itself and 2:46you'll see it eventually lay out a 2:47really clear methodology for how it 2:49solves the problem if you give that same 2:51problem to 40 it will not work nearly as 2:54well in fact I did that last night and 2:56what I saw was clearly an actual 2:59rational response from 01 and clearly 3:01much more simple next token prediction 3:04from 3:0540 and so I think what the open ai's vpa 3:10product said got it right at the end of 3:11the day the question of whether an llm 3:13is hitting a diminishing rate of return 3:16on training on data is 3:19different than whether or not you are 3:24getting more value out of reasoning and 3:26open AI has been telegraphing for months 3:29that they a lot of the value in scaling 3:31intelligence as scaling at the time of 3:34inference or scaling 3:35reasoning so we will see but it just 3:39looks like the information got very very 3:41excited about reporting something that 3:43they thought was scandalous and 3:45surprising and rushed it to print a 3:47little bit too much because it gave the 3:49impression that open AI was out of juice 3:51and the reality is open AI is just at 3:54the beginning of releasing something 3:56interesting and I actually think that is 3:58one of the things that makes this leak 4:00seem a little bit malicious 01 the full 4:0401 model is rumored to be released this 4:06week now I've been doing this for a bit 4:09I've been seeing these rumors come out 4:11so that's not that surprising that 4:13they're rumoring it this week maybe it 4:16will happen this week maybe not but it's 4:19supposed to happen by the end of the 4:20year we're coming up into the teeth of 4:21the end of the 4:23year and it is very interesting that 4:27they would choose to play for clicks in 4:29this way right now all right I've said 4:32enough about that number two the US has 4:36ordered 4:37tsmc to Halt export of roughly speaking 4:41AI chips to China if you want to dig 4:43into the details it's uh chips below the 4:457 nanometer architecture specifically 4:48going for gpus used in AI 4:51applications and you might wonder I 4:54wondered how on Earth is the United 4:56States government able to restrict 4:58exports from Taiwan to China it doesn't 5:01make sense well it turns out that we 5:04have a special set of rules called the E 5:07the export 5:09administrative something I'm going to 5:11have to look this up a second it's 5:14the uh 5:16export 5:18administrative rules that are authorized 5:22by the export control Reform Act of 2018 5:24I did my research on this um and what 5:28basically what it does 5:30is it allows you to if you are the US 5:35government regulate the export from 5:38other countries of things that are 5:39produced using US tech and in this case 5:42the chips have us design input a lot of 5:46the sort of routine in tsmc is you get 5:48the chip designed from the US and then 5:50you manufacture it in Taiwan on that 5:53basis of the intellectual property and 5:55the chips the Chamber of com the US 5:58Department of Commerce is a able to 6:00regulate export of the chips from Taiwan 6:01and that's what they're doing and the 6:03reason they did it and this is not as 6:04widely 6:05reported is that a chip that had gone to 6:09a Chinese customer which was legal at 6:11the time ended up in a huahe device 6:16which is not legal because the US has 6:18put huah on a export restrictions list 6:22and so basically seeing that like 6:24there's some unlicensed transfer of 6:27these chips inside China the Department 6:30of Commerce basically said no just stop 6:32exporting to China until we figure this 6:34out and so that's what as of today tsmc 6:38is doing now 6:40separately people are wondering how 6:42tariffs if they come will affect AI chip 6:47production and one of the plays for a 6:49lot of other materials is that people 6:51ask themselves can we do final assembly 6:53in the 6:54US in this case Taiwan has sent a very 6:58clear signal that you won't be able to 7:00because it is said explicitly that they 7:01have regulations on the books in Taiwan 7:04that say that advanced chip 7:06manufacturing capabilities in this case 7:08the current top generation is 2 7:11nanometer that they will not be able to 7:13make them elsewhere because they have 7:15regulations saying that capability stays 7:17on Taiwanese soil so we will see we have 7:20some competing 7:21regulations and in the meantime I'm 7:25hoping that these weird photos that open 7:26AI employees continue to post about the 7:29Orion Nebula end up meaning something 7:32right like surely they can't just post 7:34pictures of the stars and not release 7:35models we will see