Learning Library

← Back to Library

Kimi K2: Hype, Benchmarks, and AI Trends

Key Points

  • The episode opens with a round‑table of AI experts who debate whether the new open‑source model Kimi K2 is over‑hyped or under‑hyped, noting that while benchmark scores look impressive, its real‑world generalization remains unproven.
  • Kimi K2, launched by the Alibaba‑backed startup Moonshot, claims to surpass Claude and GPT‑4 on coding benchmarks, sparking excitement that an open‑source model can now compete with industry giants in specialized tasks.
  • The hosts caution that benchmark victories alone don’t guarantee broader utility, emphasizing the need to see how the model performs in diverse, production‑level scenarios.
  • The conversation also touches on broader AI infrastructure trends, including Google’s massive new data‑center investment and Lawrence Livermore National Laboratory’s recent shift toward cloud‑based AI workloads.
  • A brief retrospective on the “R1” project wraps up the episode, highlighting lessons learned from earlier AI initiatives and how they inform current developments.

Sections

Full Transcript

# Kimi K2: Hype, Benchmarks, and AI Trends **Source:** [https://www.youtube.com/watch?v=hNvbeXus-pM](https://www.youtube.com/watch?v=hNvbeXus-pM) **Duration:** 00:47:40 ## Summary - The episode opens with a round‑table of AI experts who debate whether the new open‑source model Kimi K2 is over‑hyped or under‑hyped, noting that while benchmark scores look impressive, its real‑world generalization remains unproven. - Kimi K2, launched by the Alibaba‑backed startup Moonshot, claims to surpass Claude and GPT‑4 on coding benchmarks, sparking excitement that an open‑source model can now compete with industry giants in specialized tasks. - The hosts caution that benchmark victories alone don’t guarantee broader utility, emphasizing the need to see how the model performs in diverse, production‑level scenarios. - The conversation also touches on broader AI infrastructure trends, including Google’s massive new data‑center investment and Lawrence Livermore National Laboratory’s recent shift toward cloud‑based AI workloads. - A brief retrospective on the “R1” project wraps up the episode, highlighting lessons learned from earlier AI initiatives and how they inform current developments. ## Sections - [00:00:00](https://www.youtube.com/watch?v=hNvbeXus-pM&t=0s) **AI Podcast Discusses Kimi K2** - In the opening of the “Mixture of Experts” podcast, host Tim Hwang and guests preview AI topics—including a Google data‑center investment and Lawrence Livermore’s cloud adoption—before debating whether the Kimi K2 model is over‑ or under‑hyped. - [00:03:17](https://www.youtube.com/watch?v=hNvbeXus-pM&t=197s) **Cautious Optimism on New Model** - Participants debate the hype versus actual performance of a new open‑source AI model, stressing the need for real‑world benchmarks and acknowledging it still falls short of top proprietary systems like Claude and GPT‑4. - [00:06:23](https://www.youtube.com/watch?v=hNvbeXus-pM&t=383s) **Open‑Source vs Proprietary LLM Cost Debate** - Participants debate the performance hype of a new model, contrasting it with proprietary APIs and highlighting shifting cost dynamics from per‑token fees to predictable compute expenses for business use. - [00:09:26](https://www.youtube.com/watch?v=hNvbeXus-pM&t=566s) **Challenges for New Coding Models** - The discussion highlights how the entrenched install base and vendor lock‑in of existing proprietary AI coding models create economic and strategic barriers that prevent even superior new models from gaining market traction. - [00:12:36](https://www.youtube.com/watch?v=hNvbeXus-pM&t=756s) **OpenAI Delays Open-Source Model** - The speakers discuss OpenAI’s indefinite postponement of its much‑hyped open‑weight model, question why major U.S. AI firms are lagging behind Chinese open‑source efforts, and reference competing models such as DeepSeek, Kimi K2, and Mistral. - [00:15:38](https://www.youtube.com/watch?v=hNvbeXus-pM&t=938s) **Open‑Source AI Model Challenges** - The speaker discusses the limited emergence of competitive open‑source AI models despite market interest, notes DeepSeek’s current lead, and probes whether developing open‑source models requires a distinct discipline compared to closed‑source approaches. - [00:18:43](https://www.youtube.com/watch?v=hNvbeXus-pM&t=1123s) **Regulatory Barriers Shape AI Adoption** - The speaker explains that compliance and tooling hurdles limit R1’s enterprise uptake in the West, while Chinese startups and academic researchers drive its use, amid geopolitical tensions and efforts to circumvent U.S. controls. - [00:21:52](https://www.youtube.com/watch?v=hNvbeXus-pM&t=1312s) **Shift Toward Safe, Efficient AI** - The speaker describes a strategic pivot from pure model capability promotion to emphasizing safety, governance, and efficiency—positioning Western AI firms as trustworthy alternatives and questioning whether U.S. companies can adapt to resource‑lean, open‑source model development. - [00:25:10](https://www.youtube.com/watch?v=hNvbeXus-pM&t=1510s) **Patrick Mahomes Effect on AI** - The speaker argues that DeepSeek‑V3's openness fuels creative AI progress, yet current models still fall short of the extraordinary, “Mahomes‑level” breakthroughs that would truly astonish users. - [00:28:18](https://www.youtube.com/watch?v=hNvbeXus-pM&t=1698s) **Google’s $25B Energy Investment Overview** - The speakers note a 17% NVIDIA stock drop yet persistent GPU demand for large‑scale inference, then shift to highlight Google's $25 billion investment in Pennsylvania hydropower and the PJM interconnect grid. - [00:31:28](https://www.youtube.com/watch?v=hNvbeXus-pM&t=1888s) **Balancing Data Center Growth with Grid Limits** - The speaker critiques Google's holistic energy approach, stressing that expanding data centers will strain the power grid, require diversified sources like nuclear and hydropower, and could raise electricity costs for consumers, particularly in underserved Midwestern regions. - [00:34:39](https://www.youtube.com/watch?v=hNvbeXus-pM&t=2079s) **Google Drives Industrial-Scale Renewable Energy** - The speaker highlights how soaring AI energy demand and long power‑connection delays are forcing big tech like Google to become major renewable energy buyers and investors, spurring an industrial‑scale race in clean power infrastructure and its downstream impacts. - [00:37:48](https://www.youtube.com/watch?v=hNvbeXus-pM&t=2268s) **Environmental Concerns of AI Compute** - Speaker questions the energy and water impacts of AI data centers and muses about alternative solutions such as space‑based computing. - [00:40:51](https://www.youtube.com/watch?v=hNvbeXus-pM&t=2451s) **AI Inequality and Claude Expansion** - The speakers caution that powerful AI models may widen economic divides while announcing Anthropic’s Claude being deployed across Lawrence Livermore National Laboratory to aid thousands of scientists with complex data analysis and hypothesis generation. - [00:43:58](https://www.youtube.com/watch?v=hNvbeXus-pM&t=2638s) **From Tools to AI Collaborators** - The speaker admires emerging AI for scientific discovery, debates whether it remains a supportive desk tool or evolves into a fully autonomous research partner, expresses concern over hallucinations, and envisions future AI agents acting as co‑authors that generate hypotheses. - [00:47:08](https://www.youtube.com/watch?v=hNvbeXus-pM&t=2828s) **Behind the Research Narrative** - The hosts discuss overlooked aspects of scientific research, acknowledge the contributors, and close the episode with thanks and a podcast promotion. ## Full Transcript
0:00It's all great in theory, but 0:02then you know what happens when it comes to. 0:06I've got a power Google AI overviews. 0:09Or are Mr. and Mrs. Jones down the road? 0:13Need to watch the television this evening or need to keep warm in the winter? 0:17And you're like, uh, I'm paying for the data center. 0:20Sorry, grandma. We have a pre-training run. 0:23All that and more on today's Mixture of Experts. 0:32I'm Tim Hwang, and welcome to Mixture of Experts. 0:35Each week, MoE brings together 0:37a crack team of the most brilliant and entertaining 0:39researchers, product leaders, and more to distill down and chart a path 0:42through the ever more complex landscape of artificial intelligence. Today, 0:46I'm joined by Abraham Daniels, Senior Technical 0:48Product Manager for Granite. 0:50Kaoutar El Maghraoui, Principal Research Scientist and Manager for Hybrid 0:54AI Cloud, and Chris Hay, Distinguished Engineer. 0:57We have a packed episode today we're going to talk about 1:00a little bit of a retrospective for R1. 1:03We'll talk about a huge data center investment by Google. 1:06We'll talk about the adoption 1:08of cloud by Lawrence Livermore National Laboratory. 1:11But today I actually want to start first with Kimi K2. 1:19And I think for our round the horn question, 1:21we'll do a really simple one, which is Kimi K2. 1:23Is it over hyped or under hyped? Uh, 1:26Abraham, curious. Have you got any thoughts on that? 1:28Honestly, I don't know. It's, 1:29uh, from a benchmark perspective, 1:31it looks amazing, but I think we have to wait and see. 1:33From a generalization perspective, it's actually as good as they say. 1:37All right, Chris, what do you think? 1:38It is actually really good. 1:40But it's not better than Claude. 1:41No matter what the benchmarks say. 1:43All right. 1:44And finally, last but not least, Kaoutar. What do you think? Yeah, 1:47I think it's a little overhyped, 1:48but yes, it's a it's a very good model. Okay. 1:51A lot to get in here too. 1:53I love these opinions. They're like ah, 1:55maybe good, maybe bad. Um, 1:57so just give a quick background 1:59for folks who may have not been watching this. 2:01So, Kimi K2 is a new model 2:03that dropped from the Alibaba backed startup moonshot. 2:08And it's an open source model notably. 2:11And it's been kind of really storming the charts. 2:13There's been a lot of chatter about it online. 2:15People are saying it's the best thing since sliced bread. 2:18And I think the most interesting thing about the launch 2:21is that the Moonshot 2:23company has basically claimed 2:25that against benchmarks, it is surpassing 2:28the latest state of the art for Claude and GPT-4, 2:31particularly on coding benchmarks, which is a big deal, right? 2:35The idea that on this specialist task of coding this open 2:38source model is now challenging, 2:41you know, the biggest players in the game. 2:43And Abraham, maybe I'll start with you 2:45because I thought your response was maybe a good way into this discussion. 2:48You were saying, well, hey, it looks great, 2:49but we actually don't know yet if it's any better. 2:53Um, what do you mean by that? 2:54Tell us. Tell us more. Well, 2:55a couple of things. One, in public benchmarks, uh, 2:58you know, as we've spoken, a couple in a number of these, uh, 3:01you know, Mixture of Experts, episodes can be gained 3:04and they don't always tell the full story. Um, 3:06so although, you know, 3:08they may have published that they're better than Claude and GPT 3:12until we can actually get some independent 3:14or third party or see what the community actually thinks. 3:17think it's, you know, maybe the claim 3:18is a little bigger than it really is. 3:20Um, also, it's, 3:23uh, I don't know my opinion 3:25that there's a lot of kind of craze at the beginning, 3:27and then things kind of settle down 3:29and we kind of figure out where it really stands. 3:30So I'm cautiously optimistic about its performance, 3:34but I'd like to just see some, uh, real world applications, 3:38whether that's, you know, integrating to certain stacks or, you know, 3:40actually demonstrating side by side comparisons, 3:43whether this is actually as good as they say it is. Yeah. For sure. 3:46And Chris, I think maybe I'll turn to you next. 3:48Like, you know, I think the caution is well warranted. 3:50And I think at this point I like barely look at like the benchmarks 3:54in the blog post when they announce models because I'm like, 3:57ah, it's all a gamble. It's all a trash. 3:59But you seem to be convinced you're like like just on playing around with it 4:02It's a good model, but it is definitely not 4:04as good as Claude and GPT-4. What, 4:06what leads you to say that, um, 4:09putting my hands on the keyboard and typing stuff in and see what comes out? 4:13Give me more than that, though. Of course. 4:14But this is more than just a vibe check, right? 4:15You actually think against certain tasks. 4:17You think that, like, still, 4:19this is not surpassing the state of the art here. 4:22No, I don't think. I don't think so. 4:23So the first thing I would say it is by far, 4:27in my humble opinion, 4:29it is the best open source model out there at the moment. 4:32Or open model. 4:34Um, it they have done a phenomenal job. 4:36I mean, it's a 1 trillion parameter model. 4:38So this thing is big, okay. 4:40It is a mixture of expert model with a lot of models, 4:43but it's still a big model. 4:44And you know, and you need a lot of disk space 4:47to get that running on your machine. 4:49Now. 4:49Um, but it is the best model for an open source, 4:53but it doesn't be closed. 4:54And there are a lot of things 4:57that I think are really good for this model. 4:59I mean, when I was playing with it, I really liked its planning capability. 5:03I really liked its tool use. 5:06So this is a model that is definitely being designed 5:10for a genetic behavior, right? 5:12They've really focused on the plan and they've really focused on the use of tools. 5:17Um, and I think that is going to be exciting 5:20when we run a smaller model 5:22because to be honest, when you want to run agents, 5:25I said agents, of course, but when you want to run agents, 5:28you want your models to be small and fast and lean 5:31and and I think it's going to do a phenomenal job as well. 5:33The other thing is it's not a reasoning model. 5:36So it doesn't have that thinking capability yet. 5:38They've just provided a base model and a model. Um, 5:41but it is fabulous for the chat. 5:43So code wise or to sort of 5:46come back to what I said, there are ten, right? 5:48Code wise I think open source, open weight model, 5:52It is the best coding model out there. 5:54I've used pretty much every single one of these models, 5:58whether it's the Qwen models, whether it's a DeepSeek, etc. 6:02it really is the the best coding model out there for an open model, 6:06but it doesn't be closed, right? 6:08It may be that on the benchmarks. 6:09And back to Abraham's point, right. 6:11is a lot of these things are gained 6:14towards the benchmarks to try and get that sort of edge. 6:16But when you put it in real coding scenarios, right. 6:19Um, I want to code up this, 6:21this particular program, change this, do this or whatever. 6:23It does a good job. 6:25But the code is better, right? 6:26I mean, cloud is giving me better results 6:29than I'm seeing from my vibe checks. 6:31Um, but but fair play to them. 6:33I don't take anything against that 6:36It is an incredible model. 6:37And for kind of the budget, the compute, 6:40the time that they've had again, spectacular. Kaoutar, 6:43I know you came in basically saying that you felt like 6:46it was a little bit of an over hyped launch. Um, 6:48and so do you kind of buy Chris? 6:50You're basically like very good, but still like, you know, 6:54as compared to the proprietors. 6:56You know, I think it's still still lagging a little bit behind. 6:59Yeah. So but I think, you know, 7:01um, there are also other angles that, you know, this, 7:04um, release or this launch is kind of, 7:07uh, getting us to start thinking about 7:10which is more on this evolving war on the cost, 7:12you know, the open source versus, you know, the proprietary APIs. 7:17So if you look at companies like OpenAI 7:19or, you know, Anthropic or Google, 7:21they're charging per token for API access. But, 7:24you know, these open source models with models 7:27like Kimi K2, Llama or Mistral, 7:29you know, the cost here is shifting from these API 7:32free to a fixed or at least predictable infrastructure cost, 7:36you know, so you're paying more for the compute. 7:38it's like we're getting, you know, 7:40with models like, you know, it is a great model. 7:42I played with it a little bit, uh, but it's kind of where 7:46the good enough tipping points. 7:48So for, you know, many business tasks, 7:51you know, you know, summarization, classification and etc. 7:54these open models are doing a pretty good job, 7:57you know, even superior than, you know, the closed ones. 8:00So now I think we're kind of getting into this phase 8:03where companies, you know, can now adopt, 8:06you know, this hybrid strategy, use maybe expensive 8:09proprietary models for maybe complex frontier tasks, 8:13but then offload the bulk of their workload 8:16to really cheaper or self-hosted open source models. 8:19So but I also feel, you know, with, you know, this launch, 8:23we're kind of getting into this maturation of the open source AI movement. 8:28I mean, it just didn't happen with Kimi K2, 8:31but also with the other, you know, open source models. 8:33So it's no longer about, you know, providing a free alternative, 8:36but also about competing directly on 8:38performance and features with these, 8:41you know, other, you know, closed source models. 8:43So I think with this release, it's also kind of pushing for, you know, 8:48kind of, um, towards putting more pressure on the pricing models 8:52of these proprietary giants like OpenAI and Google 8:55and, you know, kind of, you know, putting a lot of pressure on them. 8:59So the the future of these, you know, enterprise 9:01AI is not just a single vendor solution. 9:03I think it's kind of leaning toward more cups, 9:06optimized portfolios, hybrid models. And, 9:10you know, I think Kimi K2's success 9:13or, you know, really great performance signals that, you know, 9:16the primary battleground in AI here is shifting 9:19from this pure performance race to kind of a war of economic efficiency. 9:24And also like the strategic control. Yeah. 9:26And I think I did want to pick up on the strategic control point. Um, 9:29there's an interesting observation that some people are making, which is, okay, 9:32I know this group is maybe a little skeptical about sort of K2's 9:36ultimate capabilities on coding, 9:38but like, assume for a moment that it is actually better 9:41than what Claude and like, say, OpenAI can provide. 9:45A lot of people were pointing out that actually now it's actually 9:47a little bit difficult for Kimi to compete in that universe, 9:50because a lot of people are on platforms and endpoints 9:54that are using all the existing 9:57leading proprietary models. 9:58And I guess, Abraham, maybe I'll throw it to you 10:00because I know you're working with Granite day in and day out. 10:02Like, do you think that there's kind of this really interesting dynamic emerging where 10:06now the kind of like preexisting install base effectively, 10:10if you will, for these models, particularly in coding 10:13like means that it's actually really difficult for like a new model, 10:16even if it's better to get in and actually compete with these proprietors. 10:19Do you buy that at all? I'm not really. 10:21I think it's less about whether if it's better or not. 10:24And to to Kaoutar's point, it's really 10:26what are the economics of using this model versus, 10:29you know, vendor lock in or locking into a particular stack or infrastructure. 10:33I think the question really, is it good enough 10:35where the price tag aligns with our, 10:38you know, our business case or, you know, our user base? 10:41And I think you're consistently seeing that these, you know, 10:44open source is now a strategic weapon as opposed to just a, um, 10:48you know, a mandate by an organization 10:51where you're starting to disrupt a lot of these, you know, 10:53closed source models, and when you can actually brush up against their performance, whether that's, 10:57you know, R1 on reasoning or Kimi K2 11:00or communicate to on on coding, 11:02you're really signaling to the market that, you know, um, 11:05one vendor lock, in 11:08my opinion, of Interlochen, was always going 11:10to be kind of dismantled as you had a proliferation of developers. 11:13was going to kind of erase to the bottom, 11:15but this really just kind of expedites it. 11:17Uh, and then too, I think there's also, you know, 11:19developer centric pricing is going to continue to force a downward pressure. 11:22I think over the last six months, you've seen an actual explosion 11:24in cost per input and output tokens. 11:27So I be personally, I think this is amazing. Granite, 11:30as a, you know, as a model, 11:32we are huge proponents of open source 11:35licensing with, you know, no, not all things open sourcing. 11:37So I think this is the right direction, not only for the field. Um, 11:40and then I also think this is kind of signaling 11:43to, Llama and OpenAI that they have to start 11:46to take this very seriously in terms of how this takes into their roadmap, too. 11:49So with OpenAI starting to 11:51or hinting at another open source model, the first since GPT-2. 11:55So firstly, um, you know, back to your question, 11:58I yeah, I don't think this is necessarily an issue. 12:01I think this is really just a it's an economics question, more 12:04so than, uh, a technology question. 12:10The second topic of today that I really wanted to get into 12:13was zooming out from Kimi K2, right. 12:16Someone pointed out to me recently where 12:18six months since the R1 launch, 12:20and which is amazing because R1 launched January 20th, 2025, 12:24it already feels like it was six years ago, not just six months ago. Um, 12:28but I think it might be good for us to kind of 12:30just talk for a few minutes, zooming back a little bit on 12:33like what has changed since R1 launched. 12:36Um, and I think Abraham, you're picking up, 12:38I think, on one thing that I did want to bring up, 12:39which is, you know, in the midst 12:41of all this, OpenAI announced 12:44that it would be kind of delaying indefinitely 12:47the launch of its open source model, which was kind of way 12:49hyped and was originally read as kind of a response 12:52to this new generation of Chinese open source models, 12:54but now appears to be kind of like on the on the back burner 12:58Well, back burner is maybe the wrong word, 13:00but delayed for an unknown amount of time. Um, 13:02I guess, Chris, maybe to throw it to you like, 13:05you know, do you feel like the US companies in 13:06some ways have, like, not been able to kind of like, 13:09answer this open source challenge at all? Right. Like, 13:11I think in some ways like Meta is still competing, 13:14but OpenAI is not really open sourcing. 13:16It feels like there hasn't been another kind of marquee model that says, okay, actually, 13:20a lot of these kind of dominant US companies 13:22can kind of keep up in this race. 13:24I think there's different 13:26economics and power shifts in play in this sense. 13:29I don't think there's any reason why OpenAI or Anthropic 13:35can't release an open weight model. Right? 13:37Um, they're obviously choosing to do other things there. 13:41I stick by my statement that I said earlier for size. 13:45I think the best open weight 13:47models out there are the deep six. 13:49That is the um, you know, now surprised by became, 13:52you know, suppressed by the Kimi K2 model. 13:55Um, you know, the Mistral models are incredible. 13:59Their open weight models, um, you know, 14:01especially their 24 billion parameter one and the Mistral medium, 14:05they they're really great models. Um, 14:07and I love what we're doing with Granite 14:09with the 7B models or in the 8B models. Right. 14:12Everybody. And the, the the 1B models, 14:14I think everybody's forgetting about these really small models.. 14:17And and actually they become super important especially for things like agents. 14:21So I, I think they're missing a trick. 14:25I mean, the only American company 14:27that's really producing good open weight 14:29models is Google it that, you know, 14:31Google at the moment and, and IBM obviously. 14:34But I mean, on the kind of the higher number of parameters. 14:37Um, so I just think 14:39I think there is more to do in that effort. 14:42And and that's the you know, 14:44it's running away from there. 14:46So I'd like to see that position change. 14:49Um, because the reality is 14:51there's a risk for all these companies, 14:53which is once you start to get, um, 14:57competitive models 14:58and you're not going to compete with a trillion parameter model 15:00But if you can get a really great coding model 15:03down to the 8 billion parameter number and and and again, 15:06I don't think that's far off when you think about some of the like, 15:09Mistal's doing with the 24 billion parameters 15:11then are about to counter his point about cost economics. 15:16If I can run something on my laptop 15:17and I can get good code from it, or I can run 15:19good agents from it, that starts to affect their business model. 15:23So I, I'm a big fan of open weight, a big fan of open source. 15:27I really like to see all the close, uh, 15:29source providers open up their models and open up their weights. 15:31I'd like to see that as, uh, just get it done. Yeah. 15:34For sure. Well, 15:35and I think that's that's one thing 15:37I did want to get to in and counter. 15:38I guess you've been name checked, so I'll kind of bring the conversation 15:40back to you is like, you know, 15:42there's obviously different economics and the kind of sort of us 15:45leading companies are like trying a couple different things in the space. 15:50Um, but it is kind of interesting to me 15:52that it feels like the number of kind of like, 15:54like, I guess when I think about open source, I think like, 15:57oh, well, there's going to be tons and tons of different players 15:59putting out lots and lots of different models. 16:02And, you know, we're going to see this space really, really kind of open up. 16:05I mean, to Chris's point, you know, even though the Chinese market has kind of 16:08like really invested in open source, it still feels like after six months, 16:12DeepSeek is really still kind of like in the lead here. Right. 16:15Like that actually, we haven't seen like an explosion of new companies 16:18offering open source models in the space that are kind of 16:20at least as competitive. 16:22I guess the question I kind of want to get you to respond to 16:25is like whether or not you think there's like a special discipline 16:28with doing open source models that's maybe different from closed source. 16:31Like, is there like a different style of what's going on here 16:34that actually is almost as difficult as doing a closed sourced model 16:37well. Yeah, that's a very good question, I think. Um, 16:40what really, you know, 16:41I think help seek is the efficiency 16:44or, you know, aspect of it. 16:46So I think the key innovation was mostly behind 16:49their architectural efficiency 16:51where they employed, you know, the, you know, the bag of techniques 16:54of mixture of experts, reinforcement learning, 16:56you know, optimizations all the way to the level of the PTA, etc.. 17:00So that was I didn't think that was kind of a 17:04maybe, uh, a kind of a breakthrough thing, but more, 17:09you know, efficient implementations 17:11and clever ways of using existing techniques. 17:14So, and, uh, 17:16and, of course, you know, there is an ongoing debate 17:18about the nature of, you know, Deep Six achievement. 17:21You know, while, you know, some of them view that, you know, 17:23their methods are revolutionary breakthroughs. 17:26You know, I'm more, you know, along the lines of those that you know, 17:29that think that, you know, it's a clever and effective implementations of existing techniques 17:33rather than a fundamental paradigm shift. So. 17:36So but that was I think the efficiency was very important because, you know, 17:41showing that you can get, you know, 17:43to these state of the art models with less costs. 17:47That was I think, a very important shift. 17:50You know, that they have showcased here. 17:52Um, and since then we've seen, 17:54you know, many releases where they kept improving their models. 17:58So they have the steady, you know, flow of releases where they kept improving. 18:02So that that was really great to see. 18:04Um, so going to your question, 18:07what's kind of the maybe the the recipe here? 18:11I think, uh, of course, you know, 18:13being able to, to be state of the art 18:15kind of beating these benchmarks, 18:18but also having the capability to do these things efficiently. 18:22Um, but if you see like six months, 18:25you know, from their launch, 18:27um, have they kind of shaken the markets? 18:30Uh, have they kind of, you know, 18:33uh, like, especially the closed source ones. 18:36Uh, probably not that much. 18:37You know, the enterprise uptake for, 18:40you know, deep sky field, it still remains a limited. 18:43And I think it's mostly due to regulatory and compliance 18:47and some of the tooling blockers. 18:49So the adoption, of course, 18:51uh, is mostly concentrated in Chinese 18:54based startups and hobbyist communities. 18:56But then, you know, if you look at, you know, 18:59in the, the West and the, in the US and, you know, so the, the, 19:02the enterprise uptake is still limited. 19:05Um, and it's mostly, I think, 19:07of course, in the academic and the specialized domain, 19:10there is a lot of traction here. 19:11Like a lot of researchers are leveraging R1 19:14for math, problem solving, code generation 19:17And, you know, especially for the Chinese language, medical diagnostics, for example. 19:21But then, you know, in the enterprise, 19:23I think it's I feel it's still limited. And, 19:26uh, maybe that's also kind of part 19:30of this geopolitical, geopolitical AI 19:32race where we've seen it is getting intensified. 19:35So because deep seeks open source 19:38strategies, encouraging, you know, the rivals for example, 19:41moonshot are like we're seeing with Kimi K2 to follow here 19:45and uh, and especially to kind of trying to partially bypass 19:49the US controls or the US chip controls, 19:52so that that is really something that is so important for them. 19:56Um, but you know, what we see also on the Western governments, 20:00they're really trying to double down on these trustworthy AI frameworks, 20:04uh, which is becoming very important. 20:07Yeah. I think this is so a lot to unpack there. 20:09And I think you're getting to something I think is really interesting is 20:12I think the narrative when R1 launched was, 20:15oh, man, all of these American companies are suddenly in trouble 20:18because you have this incredibly powerful model 20:21and it's available for for free, right. 20:24And I think six months on, my kind of reflection is caps 20:27are the same as yours, which is actually like enterprise 20:29adoption has been kind of less than I would have thought. 20:33Um, and and that's, that's pretty interesting, right? 20:36That like, in some ways, the market dynamics 20:39that we kind of originally thought with R1 20:41and particularly around open source 20:43don't necessarily seem to be playing out the way we thought. 20:46I guess, Abraham, do you have any responses to that? 20:47Like, it's it's kind of odd to me 20:49that, like, you have this incredibly great model 20:51that's like available for free 20:54and we just haven't seen like mass adoption in a six month period. 20:57Like, if anything, you know, the proprietary, 21:00like your open eyes or anthropic of the world seem, 21:03you know, they're changing the strategy, but they're not like 21:05completely demolished as a result of this change. 21:07Yeah. And I think that's exactly it. 21:08I think it was less of like, uh, like competition with respect to, 21:13you know, another model that's in the queue 21:15in terms of, you know, what your enterprise is going to use. 21:17I think it was just more of like a what the strategy was. 21:21The status quo pre R1 shifted to be able 21:24to differentiate from R1. 21:26So you know where those models were clearly ahead, 21:30you know open weight was able to give you 21:32parity on key resource key reasoning tasks. 21:34So it shifted to you know let's get smarter 21:37cheaper inference as the goal. Uh, 21:39you know, agents were agent tech orchestration was already 21:42kind of, you know, bubbling up, but everybody doubled down on, 21:46you know, being able to develop 21:48an LLM that was, you know, 21:50a key supporter of agent tech workflows. 21:52Safety was also doubled down to in terms of, 21:55you know, our models are, you know, safe from a, 21:57you know, red teaming from a governance from an AI perspective 22:00both on the model and the data side. 22:02So I think it was really just a, 22:04um, a shift in strategy from like a, 22:07uh, a model capability, 22:09PR, if you will, perspective in order 22:11to differentiate from our one, 22:13to showcase that, you know, we are moving forward as a, 22:16you know, US based or Western model developer companies, 22:20um, and less of a, 22:22you know, R1 was now considered a viable option 22:25as part of like an enterprise use case. 22:27That's really interesting. Chris, 22:28maybe a final comment. Again, 22:29pulling out of kind of Qatars theme. 22:32You know, I think, Kaoutar, you pointed out, I think it's a really interesting thing, 22:35which is, well, maybe part of Our one's genius 22:38is its dedication to efficiency, right. 22:41They were able to assemble all these hacks together to really get like, squeeze 22:44a lot of results without having a whole lot of resources. 22:48Um, and I think a little bit about, like, what it means 22:50to be efficiency minded 22:52and how it can be really hard 22:54to kind of like, think in that style 22:57if you're used to having like, the most compute 23:00and the most money in the entire world. 23:02Um, and I guess, Chris, I don't know 23:04if there's almost kind of a thesis here 23:06that I want to run by you, which is like, 23:09could it be hard for American companies to pivot into this? 23:12Which is a big deal if you think that, like, small open 23:15source models are going to be like the future of agents. 23:17Is it hard for these companies to pivot into this kind of efficiency mindset? 23:21Because in some ways, technically, I think they're like 23:23maybe so used to an environment where it's like, 23:25we never have to think about 23:27how to assemble all these things to squeeze 23:29the most results out of limited resources. 23:32Um, I'm curious about if you think that's almost like a barrier 23:34in some ways to these companies pivoting towards open source. 23:37I think that when you are limited by your resources, 23:43you become super creative. 23:45And actually, if we think about the Kimi K2 scenario, 23:49they got super creative, right? 23:51One of the biggest things that they did is 23:53they came up with their, um, their 23:55new optimizer. Right? The, 23:57the Muon optimizer, which, 23:59which was really about them being able to, um, 24:04train very, very large models in a consistent way, um, 24:08and not, uh, 24:10basically have their training losses mess up during that process. 24:15That is a huge moment. Now, 24:17we don't know all the details behind that, 24:20but the innovation there is great. 24:23They've moved away from the optimizers others are using, right? 24:27When I think about the deep seek moment 24:29and their efficiency, they similarly. 24:32But nobody really cared about deep seek. 24:34But when they first launched anyway, DeepSeek-V3 came out in December. 24:38But it wasn't until they released our one where we got excited, 24:41and it's because they had, um, 24:44the reasoning model, and it was pretty much close 24:47to the old series of models there. Right? 24:49And then they were open about how they published it. 24:52They went through, um, 24:54you know, their RL flow 24:56and how they train the, the grp stuff, etc. and, 24:59and we all learned stuff and it was all great, but they were innovative and, 25:03and the great thing is they were open about it 25:06and everybody's been running around copying their techniques and learning from them. 25:10Kimi K2 wouldn't exist if DeepSeek-V3 wasn't open 25:14about how they trained the V3 model, 25:16so I think that in itself is going to boost that creativity. 25:21But to your point, I'm not quite sure 25:24if you're just sitting there with, 25:26you know, hundreds or thousands of each one. Hundreds. 25:29I'm not sure you're 25:32you've got all the compute you need, right? 25:34I'm not sure you're going to be. 25:35So, you know, 25:37you're just going to get your job done as opposed to like going, 25:40oh, I can't do this because I don't have this 25:42and I need to figure my way out of it. 25:43So I think, I think that is helping them. 25:46But but why is deep seek 25:49maybe, you know, six months on to your point. 25:51I'm going to call it the 25:53I'm going to call it the Patrick Mahomes effect. Right. 25:55There are great quarterbacks 25:57kicking around Tom Brady who is the greatest. 26:00And then great quarterbacks who come along and you go oh there's their golf. 26:03They're like you know you're like oh okay. 26:05Even Justin Herbert people who shoot me for that they'll go ah okay. 26:09Do you know what I mean? Because you're not seeing anything amazing. 26:13Over time you get used to them. 26:14But then when you look at Patrick Mahomes play 26:17and you're like, how did he do that? 26:19No human on earth is able to make that through. 26:22How did he do? He. 26:24He wasn't even looking. 26:26And and I don't think those models are quite doing that yet. 26:30Right. Because the models that have come out 26:32are equivalent or they're, 26:35they're about the same 26:38as the, the, the models 26:39and nobody really cares about the same. Right. 26:42If you think of a like, 26:44you know, if you think of a Super Bowl or what, 26:46nobody remembers who lost the Super Bowl. 26:48They're close enough to the the team that won. 26:52Right. But but people care about the winners the greatest. Right. 26:55So I think for one of these to take hold 26:58and really upset OpenAI and throw, pick etc., 27:03they're going to have to do something like the no model has ever done before. 27:06It's just like, oh, I press a button and it's it's 27:09created an entire billion dollar company overnight. 27:12Wow. And it's done it on a chip that runs on my laptop. 27:16That will be like, whoa. 27:17I mean, they're not impressive. 27:19I would be impressed. 27:20Nobody's going to care at that point. 27:22You think you're going to stick on? 27:23I don't know, I'm going to stick typing in ChatGPT. 27:25You're like, no, I'm running over the new thing. 27:28I've got to see that. 27:29Whereas if it's just like, ah, it's the same as it was before. 27:32You're like, well, it's just the same. 27:33I'll stick with what I've got. 27:35That's what needs to change. 27:37And also I think the, the first movers 27:39adventure's always has a big, you know, effect. 27:42You know, if you know, I think OpenAI, which, uh ChatGPT, 27:45you know, kind of gained a lot of, you know, mass, 27:49uh, adoption. 27:51And so once you get used to that, you know, 27:53sometimes switching from that environment to something else, 27:57you know, you know, you really need to have like, Chris 28:00say something completely kind of a wow effect, 28:03something not just incremental. 28:05And, and I think I'm going to back to your, uh, like the, the 28:09resources or the compute question. 28:12So even, you know, you know, our one kind of shakes, 28:15you know, the, the GPU dominance, the NVIDIA GPU. 28:18So the stock dipped, you know, significantly, like 17%. 28:21But then the demand for NVIDIA hardware 28:24kind of rebounded because large scale inference still relies a lot on GPUs. 28:28So so I mean, we had the panic moments, 28:31but then but the efficiency gains, 28:34you know, really haven't negated the massive compute needs 28:37that are still, um, still there. 28:39Yeah, I think it's right. 28:40Well, we'll be checking in again another six months, I think. 28:43Kind of like using R1 as a peg and kind of moving out 28:45I think is really useful just because the space moves 28:47so, so quickly. 28:52I can move us on to our next topic. 28:55Um, announcement coming out of Pittsburgh. 28:58Really big event this week. 28:59Uh, the president was there. 29:00All the major companies were there. 29:02Um, but I think there's one announcement in particular I want to zoom in on, 29:05which is, uh, Google announced 29:07that it'd be making a $25 billion worth of be, 29:12um, announcement to invest in energy infrastructure. 29:16Uh, so for one part 29:18hydropower in Pennsylvania, 29:20and then also something that's known as the PJM interconnect, right, 29:24which is a network grid that stretches across 29:27New Jersey, Pennsylvania, West Virginia, Virginia, 29:29really large area of the country. Um, 29:32and, you know, 29:34I think this is in some ways like taking a step back, 29:36like both wild both in terms of the dollar amount being committed, 29:40but also just to remind ourselves 29:42that, like, Google is like a company that started doing search. Right. 29:46And so it's not intuitively obvious that you would eventually say years later, 29:50you know, we're going to be investing billions of dollars 29:52in going all the way upstream to really, literally change, 29:56like the energy grid of a whole part of the country. 29:59Um, and so I guess 30:01Abraham question for you is just like, 30:03how far do you think this all goes? Right. 30:05Like at some point, does Google just say we're going to be owning 30:08and operating a nuclear power plant? 30:10Like, it feels like in some ways, AI is generating such demand on the grid. 30:14These companies really need to assure energy access. 30:17And at some point, it kind of feels like, okay, 30:19where this all goes is like vertical integration. 30:21You can you can subscribe to have your energy bill sent to you from Google. 30:24Is that where this is all going? 30:26I mean, uh, it's a great question. 30:29Um, I think you can 30:32Microsoft and Meta have both, 30:37you know, committed massive amounts of money to build their own data centers. 30:40I think Google's taking a different approach in terms of not only building data center, 30:44but what I think was missing with the 30:46the prior ones is investing in the actual grid themselves, 30:50as well as investing in the community around them. 30:53Um, so to your question, 30:56you know, maybe it kind of makes sense 30:58if you talk about the actual cost of power 31:00to be able to manage these data centers. Um, 31:03if anything, I kind of, you know, clap to Google to 31:07to actually take more of a holistic approach 31:09in terms of being able to create, 31:11you know, um, compute or to create data centers, 31:15because I feel one thing that's typically missing is, 31:18you know, getting a better understanding of how what the impact of these data centers 31:21are to the surrounding, whether it's the grid, the, 31:24you know, ecosystem, the 31:25you know, this takes a ton of water to be able to cool these things. 31:28the runoff, um, so I think from Google's perspective, 31:32they did a more of a holistic approach, which I kind of applaud to you. 31:35I think this is only can we continue to happen. 31:37And you kind of mentioned, you know, nuclear energy. 31:39Like I think the next step is really better understanding, 31:43you know, whether it's hydropower, you know, electricity, nuclear is like, 31:46where's all this energy actually going to come from? 31:48Because, you know, depending on what you read, 31:50you know, by 2030, data 31:52centers are going to represent 1% to 3% of all power on the grid. 31:55And right now it just can't, you know, support 31:57that, let alone manage it. 31:59So it's it's really kind of focusing on how do we support today. 32:03And then how are these hyperscalers going to invest in the grid 32:06if they're going to be the primary user of the energy coming off of it? 32:10Um, because there are some downstream impacts. 32:12And I mentioned, you know, an environmental. 32:13But when you have all these data centers or these players, you know, 32:17integrating to the grid that drives electricity costs up for, you know, your everyday consumer 32:22and, you know, some of these areas that these grids are built by. 32:25These data centers are built in Middle America. 32:28Um, you know, these aren't areas 32:30where, you know, you typically have, 32:31uh, you know, access to as much as you would 32:34maybe like in New York or Boston or in San Francisco. 32:37So, uh, I think it's just important 32:39to kind of take a little bit more of like a long tail view in terms of, 32:42you know, 32:43building out the grid 32:45and building out these data centers and really focusing on, 32:48you know, what are the impacts above and beyond, 32:50you know, the business side of things and where the impacts from the, 32:53you know, the surrounding community and environment. 32:55Yeah. 32:56Are you think about hardware a lot. 33:00Um, and I think that, um, 33:03you know, one of the things I love about AI is how it just kind of 33:05inverts our sense of what's abundant and what's scarce, 33:09you know, like, I think a few years ago 33:10you would have said, oh, there's just so much data. 33:12We're never going to run out of data. 33:13And I think in AI land, we routinely have conversations where we're like, 33:16how do we get the next most valuable tokens? 33:19And it feels like for a long time, 33:21at least in what we're talking about here, 33:24like hardware felt like the real bottleneck, 33:26which is like, can you get access to Jensen's chips? 33:29That really was the big thing. 33:31Over the longer run, though, 33:33the midterm, like let's say 5 to 10 years, like, 33:36do you think energy becomes the new bottleneck? 33:38Like at some point I think like there will be more chips, 33:41there will be more GPUs, there'll be more suppliers of those GPUs. 33:44There'll be changes in models that maybe make the specific hardware less necessary. 33:49But it kind of feels like maybe where this is going 33:51is that whatever hardware platform 33:53you use, the energy demand is just going to be enormous 33:55And so should the world of AI start to think about 33:58like, energy becoming a bottleneck? Yeah, 34:00I totally agree. I think it's interesting 34:03to see this shift from a chip shortage to power shortage. 34:07I think, like you said, for the last few years, 34:10you know, the main bottleneck was securing enough 34:13GPUs, enough NVIDIA GPUs. 34:15But now it seems like the new bottleneck is physical security, land permits 34:19and most importantly, access 34:21to these massive amounts of stable electricity. 34:24Because the data centers, of course, is useless 34:26if you can't power and cool it. 34:28And I think even utility companies there, 34:32you know, they're reporting that requests for new data center 34:35connections are really overwhelming their capacity 34:38and forecasting capabilities. 34:39So, you know, I think wait time for, 34:42you know, large scale power 34:44connections can can be years long. 34:46So this is kind of pushing to this sustainability challenge 34:50that we're going to be facing. 34:51And I think we already started seeing these things. 34:53So this massive increase, you know, in energy demands 34:57puts enormous pressure on the climate goals here. 34:59So how do we power, you know, this AI revolution 35:03without relying on fossil fuels. 35:05So and that's what you know, Google is doing here. 35:08So I think this is forcing big tech companies 35:11to become also energy players. 35:13So they are now among, I think, the largest purchaser 35:16purchasers of these renewable energy 35:18through like the power purchase agreement, like the PPAs. 35:21And I think Google's investment here is likely, 35:25you know, tight also this new solar wind 35:28and potentially next generation 35:30geothermal or even nuclear projects to meet, you know, its, 35:34uh, carbon free energy goals. 35:36So, um, so of course, I think 35:38what Google is doing, this is a massive investment. 35:41It's just confirming that the I res right 35:43now is officially an industrial scale 35:47energy and infrastructure race. 35:49So and uh, 35:51like like you said, the the new bottlenecks right now 35:54it's going to become, uh, energy. 35:56Chris, one of the things I'm wondering if you can opine on 35:58is, I think a little bit about the like, downstream effects of all this, right? 36:02Which is you're just building a lot more energy capacity. 36:06But like the nice thing about energy is 36:07you can use it for all sorts of things, right? 36:09You could use it for industrial manufacturing. 36:12You know, there's all sorts of things 36:14that happen when energy becomes more available. 36:16And I guess I'm curious about how you like think a little bit about that. 36:20I mean, I guess maybe I'll put it in the most dramatic way. 36:23Like, if you're a cynic, you might be like, ah, 36:25all of this I stuff is a huge bubble, 36:27and at some point it's all going to fall apart. 36:30Even if that's the case, at that point, we would have built this huge 36:34electrical grid, which is kind of like this really interesting outcome 36:37is basically like it almost feels like AI is now 36:39pulling other making things happen 36:42that are going to have all these downstream effects that have like things 36:44that nothing to do with AI at all. 36:46So yeah, I don't know. 36:48Is like, I'm curious 36:50if there's like maybe to put a question on it is like if there's 36:53particular effects that you think are the most interesting here. 36:56I don't know if I'm honest. 36:58And it's not often I say, I don't know, but I imagine 37:02imagine if we went back 150 years 37:05and Google made steam trains. 37:07I'm like, do I need 100,000 steam trains? 37:11Do I need, you know, millions of tracks of Clackety Wood railways. 37:16And I and I'm like, I don't know, do you know what I mean? 37:20I, I, 37:21I cannot fall behind on building railroads. Yeah, yeah. 37:26And this is and then it would be like. 37:28And then you're like, we need more kettles to fill up the engine with water. 37:32You know what I mean? I'm not sure I, 37:35you know, like the downstream effect 37:37is like, it's all great in theory, 37:40but then, you know, what happens 37:42when it comes to I've got to power Google 37:45AI overviews or or Mr.. 37:48Mrs. Jones down the road. 37:51Need to watch the television this evening or need to keep warm in the winter. 37:54And you're like, uh, I'm paying for the data center. 37:58Sorry, grandma. We have a pre-training run. 38:00Exactly. 38:01And I so I don't really know how that works out logistic wise. 38:06And I worry about then big massive dams 38:09filled with water for the cooling 38:11and then the poor person at the other end of that dam going, 38:13I've got no water in my 38:15you know, I think there's a lot of effects and I'm just 38:18I'm not sure 38:20I'm not sure how this works. 38:21What I would like to see is people figuring out 38:24how to get more energy efficient, electricity, etc. 38:28you know how to bring down the cost of compute, 38:31you know, have more efficient models. 38:32I mean, I mean, in theory, I think it all sounds great 38:35that if if you can have the infrastructure and energy 38:37and then, um, regular people are going to as opposed to 38:41AI, people are going to get the benefit of that, then I think it's wonderful. 38:45But I, I don't know, I don't know if we're going to have 38:48like some big wasteland at the end of this. 38:51So but maybe it maybe they don't know about it all wrong. Right. 38:54Well, who says the data centers need to be on planet Earth? 38:57Why not just load it in a big rocket ship, push it towards the sun? 39:01You know what I mean? You get all the energy you want in space? 39:04And then just send the model weights down. 39:06So maybe, maybe, maybe they're doing it all wrong. 39:09I don't know. Yeah for sure. 39:10Yeah I think that's that's kind of you're getting to 39:12I think what I, what I was interested in was basically like, 39:14how much of this is really required for the future of AI? 39:18What are all the alternative structures we could imagine building? 39:21Um, but there's yeah, 39:22a lot to talk about there. I mean, maybe we'll get lucky. 39:25Maybe, maybe maybe one of these, um, compute, 39:28um, you know, like the Alibaba's, the moonshots, etc. 39:31maybe because they're so GPU constrained, they'll come up with a model 39:34that runs really small, and then then we won't need it, you know? 39:37Totally. Yeah. There's, I think, an alternative world where it's just like, 39:39actually, maybe if some of what we think is going to happen, 39:43you know, say we buy Chris your theory about like, okay, 39:45an agent world, you're going to mostly need smaller models 39:48that can kind of run locally and on devices, 39:51you know, like if that ends up being the major 39:53commercial use for this technology. 39:55What is all this huge investment for on energy infrastructure? 39:59And I think that's that's, you know, a very real outcome potentially. 40:01But maybe they're just going to be more and more usage of things. 40:05They're just going to, you know, drive, you know, more, 40:07you know, demand on the electricity. 40:09It's like right now the, the, the phones 40:12even, you know, they're kind of relatively low power 40:14But the massive, you know, 40:16usage of the phones is still going to increase the energy. 40:19So if we have all these AIs in all devices, 40:22all embedded devices everywhere, so it's still going to be, 40:25you know, I think a big energy footprint 40:28that is needed to sustain all of these things. 40:31So I think the energy problem is still going to be there, 40:34whether we go towards smaller models 40:37or we still, you know, have a hybrid approach 40:39with big and smaller models, 40:41energy is still going to be an issue. 40:43And I'm worried like, you know, Chris said, you know, 40:46what's the imbalance this is going to create? 40:48Um, the new wars. 40:51You know, they're going to be I mean, are we going to kind of increase, 40:54you know, the divide between, you know, the poor 40:56and the wealthy and, uh, accessibility to, you know, 41:00the basic things to live in favor 41:03of powering these models and things like that. 41:06So that is something I think that's a bit scary. 41:08Yeah, the movie becomes a documentary as opposed to a movie, 41:13and everybody's going to go and Google that now and go, what? 41:16That? What is Chris talking about? 41:18What is Chris talking about? 41:24All right. 41:24Last segment, which we're going to do really quickly as usual. 41:27Way more to talk about than we have time for 41:29um, fun small announcement 41:32uh, that Anthropic made on its blog. 41:35Recently they basically announced that one of their customers, 41:38Lawrence Livermore National Laboratory, 41:41one of the big national labs in the US, has decided 41:44to kind of expand their installation of Claw 41:46to expand across the entire laboratory. Right. 41:49So this is a license of their core product 41:51that goes to 10,000 scientists. 41:54So on some level, this is just, hey, 41:56you got a new customer, you got a bigger customer. 41:58That's great. 41:59But I think what's really interesting is they went a little bit 42:02into detail on what scientists 42:04at Lawrence Livermore are using Claude 4. 42:07And, you know, one of them, I'll just read it. 42:10Quote, you know, we're basically they're saying like we use 42:14or the scientists are using basically Claude for "processing and analyzing 42:17complex data sets, generating hypotheses, 42:20exploring new research directions 42:22with an AI system that understands scientific context." 42:25And the idea here is to literally use 42:27sort of a genetic, or in the very least, kind of AI assistance 42:32to accelerate scientific discovery. 42:34Um, and I guess, Abraham, maybe to throw it to you, 42:37this is like a pretty big deal. 42:38It feels like I know in the past we've talked a little bit about 42:41like, well, is AI going to accelerate science? 42:44This seems to be like a big lab saying, 42:45we're going to make a bet on this technology. 42:47Do you feel we're now kind of entering an era 42:49where AI is really going to actually be accelerating science? 42:52Um, I mean, I think it already has, to be honest. 42:55Like, I think this is just more of, like, a publicly Facing 42:58um PRP, demonstrating 43:00one of the biggest research firms in the US, 43:03if not the world, using AI to accelerate science. 43:06Um, I think what this really cool 43:07is, what I think is really cool here is, you know, the authentic 43:11kind of validated the authentic framework in a, 43:14uh, kind of like a high stakes environment, if you will. 43:17Um, but yeah, I think this is kind of, 43:19you know, an early indication of what we can do. 43:22Um, what I think is, 43:24well, I wouldn't say neat like, 43:26is there the, you know, the 43:29these are really highly like, you know, 43:32secure, uh, spaces, 43:35if you will, or like, you know, in terms of the, 43:37like, the science behind it, 43:38like having an agent and I don't know whether this is like, you know, 43:41an agent that is, uh, unmonitored 43:44or whether there's some type of human in the loop 43:46validation scheme as part of the workflows. 43:48Um, but yeah, I think look, from the perspective of using cloud 43:52and using it in a way to, you know, drive, 43:55uh, you know, scientific discovery like I want. 43:58I think that's amazing. 43:59But too, I'm also kind of cautious in terms of, you know what? 44:02You know, where is the 44:04where is it a full agentic or 44:06LLM based approach versus 44:08is this basically just like, you know, 44:10a side of a desk tool that helps navigate some of the, 44:12you know, pieces of the, of the, of the discovery or the experimentation pipeline? 44:17in short, yeah, I think this is awesome. 44:19And I think it's a sign of things to come. 44:21I think we still worry a great deal. 44:23I do at least about hallucinations all the way. 44:27These models kind of fail. 44:29Um, and, you know, 44:31I'm sure they're deploying this stuff in a responsible way, 44:34but I think the dream is ultimately what Abraham's talking about, 44:37which is you literally have an AI agent 44:39that is kind of like a research collaborator, like a coauthor, potentially on a paper. Um, 44:44how close are we to that world? 44:46I think this is kind of I feel we're entering, 44:49you know, this holy grail of generative design 44:52where we're moving from AI that analyzes to AI that hypothesis. 44:56So and of course, you know 44:58of course there are still going to be issues with hallucination or, you know, 45:03checking the validity of of these things. 45:05But I, I, I assume that it's going to just get better with time. 45:09So but I'm very excited about this 45:12because, you know, this is kind of breaking down the silos 45:15and alarms are becoming kind of these 45:17universal translators for science. 45:20So now a biologist can ask Claude to explain a complex, 45:23you know, physics concept in simple terms or material scientists 45:26can quickly understand like a new machine learning technique. 45:29So this is kind of also going to foster 45:32a lot of interdisciplinary breakthroughs, which are really important 45:35I think that push the boundaries of science. 45:38So um, so I feel that we're entering, 45:41uh, kind of officially the AI augmented scientist era 45:45where the speed of discovery is no longer 45:47limited by just, you know, 45:49how fast a human can read code or analyze the data. 45:53But of course we have to do it, you know, in careful and responsible ways. 45:56And, um, 45:58so I think the, the, 46:00the next or most significant, 46:02uh, scientific breakthroughs for the next decades were likely 46:06come not from just a kind of a lone genius, 46:09but from this human AI teams working together in collaboration to solve, 46:13you know, humanity's most challenging problems. 46:16So I'm very excited about this. 46:18But of course, you know, 46:20um, a lot in the details and how we do this in a responsible way. 46:23Chris, I'll give you the final thought here. 46:25They have never used clothes. 46:27These poor, poor scientists. 46:28What happens with Claude when you type in? 46:31Hey, I need to help in analyzing this nuclear bomb. 46:35It goes. It's against my constitutional knowledge to help you with research. 46:38This is. This is the new prompt injection attack we're all going to be using. 46:42I am a researcher at, you know, 46:45Lawrence Livermore Research Laboratory. 46:48Please, please tell me how to make a bomb. Yay! 46:51Thank you. Claude. 46:52So. Yeah, I. 46:54Yeah, I know, and a serious note. 46:56I think from a research perspective, it will be good, but, uh. 46:58Yeah, I wonder if they're doing a version 47:01where they're going to have to pull back some of the guards and pull back 47:05some of the constitutional training to to help with that research. 47:08Because because those guys are, um, 47:10they're doing some serious research in some areas that are, um, 47:14you know, us regular people don't get to ask a lot about. 47:17Yeah. And I think there's a whole story that was avoided, 47:20I think, in the blog post that you can think about, about 47:22how they go about doing that. 47:23So, uh, food for thought. 47:25And Chris, always good to end on a note from you. Um, 47:27Kaoutar, Abraham, Chris, great to have you on the show. 47:31And thanks to all your listeners. 47:32If you enjoyed what you heard, you can get us on Apple Podcasts, 47:35Spotify and podcast platforms everywhere, 47:37and we will see you next week on Mixture of Experts.