Learning Library

← Back to Library

Stargate AI Plan Premature and Exclusionary

Key Points

  • The proposed “Stargate” AI infrastructure plan prematurely declares OpenAI (backed by SoftBank and Oracle) the winner, ignoring the continued competition from Anthropic, Meta, Google, and emerging model makers.
  • Critics argue that crowning a single winner undermines the dynamic AI landscape, where numerous companies are rapidly advancing with new models, synthetic‑data generation, and innovative compute strategies.
  • The plan is based on a 2023 architecture that assumes ever‑larger GPU clusters and massive data sets, a paradigm that recent research shows yields diminishing returns and overlooks newer efficiency‑focused approaches.
  • It neglects the shift toward inference‑time compute and continual‑learning models that enable smarter AI with far less hardware, rendering the proposed massive data‑center strategy increasingly outdated.

Full Transcript

# Stargate AI Plan Premature and Exclusionary **Source:** [https://www.youtube.com/watch?v=L4TT6OAtuS0](https://www.youtube.com/watch?v=L4TT6OAtuS0) **Duration:** 00:06:19 ## Summary - The proposed “Stargate” AI infrastructure plan prematurely declares OpenAI (backed by SoftBank and Oracle) the winner, ignoring the continued competition from Anthropic, Meta, Google, and emerging model makers. - Critics argue that crowning a single winner undermines the dynamic AI landscape, where numerous companies are rapidly advancing with new models, synthetic‑data generation, and innovative compute strategies. - The plan is based on a 2023 architecture that assumes ever‑larger GPU clusters and massive data sets, a paradigm that recent research shows yields diminishing returns and overlooks newer efficiency‑focused approaches. - It neglects the shift toward inference‑time compute and continual‑learning models that enable smarter AI with far less hardware, rendering the proposed massive data‑center strategy increasingly outdated. ## Sections - [00:00:00](https://www.youtube.com/watch?v=L4TT6OAtuS0&t=0s) **Critique of AI "Stargate" Initiative** - The speaker contends that the newly announced trillion‑dollar "Stargate" AI infrastructure program prematurely crowns OpenAI as the inevitable winner, ignoring the rapidly evolving competition from Anthropic, Meta, Google, DeepSeek, xAI, and other model makers. ## Full Transcript
0:00Stargate is out it is both a pretty 0:03terrible TV show for the 1990s and also 0:07a half a trillion dollar infrastructure 0:09program that was just announced all 0:11about 0:12AI I have some real 0:14questions the issue with Stargate as far 0:17as I can tell is that it crowns a winner 0:21before the race is 0:22over it says open AI is going to win the 0:25game they're going to win it funded by 0:27Soft bank and Oracle is going to build a 0:30data centers obviously for those three 0:32players that's 0:34great even Microsoft gets in on the game 0:36they're happy to they're partner of open 0:38AI Nvidia of course is supplying the 0:41chips the problem is that there are a 0:44lot of other players in the game and it 0:47is not clear how this reshapes the race 0:49for them they're not giving up anthropic 0:51is not giving up I know I just did a 0:52video on them but they're not giving up 0:55and meta's not giving up Google's not 0:58giving up the Stakes are too 1:01high and yet here we are crowning a 1:03winner and I'm not even getting to the 1:05shifts that we've seen with model makers 1:07like deep seek entering the scene or how 1:11x. a is coming on quickly with 1:14Incredible gains uh huge compute 1:18clusters and so when I look at the 1:21problem space and I say to myself this 1:23is a hugely Dynamic situation there's 1:25lots of model makers they're all 1:28competing how does it make sense to have 1:30only one model maker get in on this 1:32project I don't think it 1:35does and I think it exists that way 1:38because this was Sam Altman shopping a 1:42deal for this kind of a data center back 1:46it feels like almost a year ago like it 1:48was like 10 11 months ago and then it 1:51died down and now it's back so that 1:55brings me to my second issue with 1:58this this is a 2:002023 2:02architecture that they are describing 2:05not a 2025 architecture and I don't know 2:08why that they are like what that doesn't 2:10make sense we've learned so much this is 2:12such a dynamic space it changes so fast 2:15so I'll explain what I mean 2:172023 we thought that we had to have ever 2:22bigger clusters of gpus to train on ever 2:26bigger data sets in order to make these 2:29mod 2:30smarter we thought that in early 2024 2:33too that's why this thing is talked 2:34about as having 10 million 2:38gpus 2:40well the thing we discovered is that at 2:45the end of the 2:46day you can have all of that compute but 2:49there may be diminishing marginal 2:50returns just for 2:52pre-training you have issues finding the 2:54data unless you're generating it 2:56synthetically which we've made some 2:58progress on but that's a big scale up in 3:01synthetic data production if that's what 3:02we 3:03do as as Ilia famously said in November 3:07of last year long after Stargate was 3:09first kind of kicked around we have one 3:11internet right like we have one internet 3:13siiz data pool we've used it 3:17so I think the reason why that feels 3:20weird in that context is that this is a 3:24architecture that fundamentally 3:26assumes this sort of older Paradigm for 3:29how trained AI models and the new 3:31paradigm the one that's unlocking 3:32continual progress that everyone's 3:34excited about it's not mentioned 3:37inference time compute is a very 3:39different Paradigm it allows you to run 3:42multiple threads simultaneously is what 3:44happens when the model thinks frankly 3:46Gemini dropped a version of that 3:48yesterday with their new update to flash 3:502.0 thinking it apparently I haven't had 3:54a chance to even try it yet apparently 3:57it's on par with o 4:01Pro 4:03so the model makers are continuing to 4:06compete they're competing on different 4:08architectural 4:09standards and Stargate is sitting here 4:12like with this 2023 structure and 4:14everyone's saying it's going to be a you 4:16know vaccine for cancer this and that 4:17well maybe 4:19but it's a weird way to go about it now 4:22and it makes me wonder if we've seen 4:24this much drift in the way we do AI in a 4:27year because we're learning so much and 4:30this thing takes four years is this just 4:32going to feel outdated by the time we're 4:34done with it it 4:36might it 4:38might and that kind of comes back to the 4:40goaling like in other major 4:43infrastructure projects that America has 4:45undertaken we've had very clear goaling 4:47you go to the Moon you bring back the 4:49astronauts it's super clear by the end 4:51of the decade they even had like a 4:52classic timeline on it 4:54fine this is not very clear it's like 4:58yeah we'll do some answer 5:00stuff okay what what does done look like 5:03what does good look like does this mean 5:06that like also the defense department 5:08will be using it maybe it's not really 5:12clear who gets to decide how all of that 5:14compute resource is allocated that's 5:17also not clear does soft Bank decide 5:19that I doubt it 5:24so I have a lot of 5:26questions as you can probably tell I 5:29think it absolutely reshapes the race 5:32it's worth talking about I put more 5:34thoughts on my substack but at the end 5:36of the day to me this is a project 5:41that makes me tilt my head and think and 5:45raises more questions that it answers 5:47and everyone's sort of talking about it 5:49as if it's a done deal it's obvious like 5:54this is it I I don't know like I don't 5:58think building the future on two years 6:01ago architecture four years from 6:04now is 6:06automatically the win maybe like maybe 6:10they just repurpose the compute I don't 6:13know but it feels a little odd what do 6:17you think