Learning Library

← Back to Library

Is Manus AI the Next DeepSeek?

Key Points

  • The panel debated whether Manus AI represents a “second DeepSeek moment,” with mixed opinions ranging from cautious optimism to outright skepticism.
  • Vyoma Gajjar highlighted the bullish case, noting Manus AI’s multi‑purpose agent could industrialize intelligence by leveraging large‑language‑model advances and potentially outpace many emerging agentic startups if hardware and compute align.
  • Kaoutar El Maghraoui expressed doubt, pointing out that numerous competing frameworks and rapid catch‑up by other firms could limit Manus AI’s long‑term impact.
  • Host Tim Hwang framed the discussion within broader AI trends—such as scaling laws, vibe coding, and new products like Perplexity’s phone—while emphasizing the significance of agentic AI as a growing focus in the industry.

Sections

Full Transcript

# Is Manus AI the Next DeepSeek? **Source:** [https://www.youtube.com/watch?v=Ddh3p185KhA](https://www.youtube.com/watch?v=Ddh3p185KhA) **Duration:** 00:49:58 ## Summary - The panel debated whether Manus AI represents a “second DeepSeek moment,” with mixed opinions ranging from cautious optimism to outright skepticism. - Vyoma Gajjar highlighted the bullish case, noting Manus AI’s multi‑purpose agent could industrialize intelligence by leveraging large‑language‑model advances and potentially outpace many emerging agentic startups if hardware and compute align. - Kaoutar El Maghraoui expressed doubt, pointing out that numerous competing frameworks and rapid catch‑up by other firms could limit Manus AI’s long‑term impact. - Host Tim Hwang framed the discussion within broader AI trends—such as scaling laws, vibe coding, and new products like Perplexity’s phone—while emphasizing the significance of agentic AI as a growing focus in the industry. ## Sections - [00:00:00](https://www.youtube.com/watch?v=Ddh3p185KhA&t=0s) **Debating Manus AI's DeepSeek Potential** - Experts debate whether Chinese startup Manus AI heralds a new DeepSeek-like breakthrough as they preview its versatile agent’s wide‑range capabilities. - [00:03:02](https://www.youtube.com/watch?v=Ddh3p185KhA&t=182s) **Skepticism Over Manus AI Impact** - The speaker doubts whether Manus will deliver a genuine breakthrough in AI autonomy or merely represent a rebranded incremental step, emphasizing the need for thorough evaluation amid intense competition between Western and Chinese AI firms. - [00:06:10](https://www.youtube.com/watch?v=Ddh3p185KhA&t=370s) **Integrating Open-Source AI Orchestration** - The speaker describes a platform that combines Claude’s orchestration with fine‑tuned Qwen models, sandboxed tool‑calling, and a cohesive UI—highlighting that while it adds value beyond a simple wrapper, the open‑source community could feasibly replicate the setup. - [00:09:20](https://www.youtube.com/watch?v=Ddh3p185KhA&t=560s) **Predicting the Future of Autonomous Agents** - Panelists examine the gap between POCs and enterprise integration, regulatory hurdles, and forecast how open‑source initiatives and projects like Manus will drive autonomous agent technology over the next six months. - [00:12:22](https://www.youtube.com/watch?v=Ddh3p185KhA&t=742s) **Agentic AI Demo Debate** - The speakers discuss the popularity of flashy, agentic AI demos as a perceived milestone toward AGI, critique current browser‑based interfaces (e.g., cursors and screenshots), and touch on experimental efforts to replicate existing agents like Manus using MCP. - [00:15:27](https://www.youtube.com/watch?v=Ddh3p185KhA&t=927s) **Debating AI's Role in Coding** - The speaker argues that although AI assistance can aid small projects, deep understanding of fundamental coding concepts remains essential for interviews, large‑scale development, and sustainable engineering practice. - [00:18:31](https://www.youtube.com/watch?v=Ddh3p185KhA&t=1111s) **Concerns Over AI‑Driven Coding Rigor** - The speaker worries that reliance on “vibe coding” with AI tools may erode code quality and best‑practice habits—especially among students—while acknowledging its potential for faster, exploratory development. - [00:21:41](https://www.youtube.com/watch?v=Ddh3p185KhA&t=1301s) **LLM-Powered Code Fixes Replace Debugging** - The speaker enthusiastically describes swapping tedious manual pointer debugging in Unix C for instant large‑language‑model corrections via copy‑paste or UI tools, while noting a worry that reliance on AI may erode deep software‑hardware co‑design skills. - [00:24:45](https://www.youtube.com/watch?v=Ddh3p185KhA&t=1485s) **Specialization and Divergence in Coding** - The speaker likens software teams to early aviators, suggesting that as machines become more reliable, developers will specialize and require less low‑level knowledge, creating a split between high‑level app builders and deep technical experts. - [00:27:47](https://www.youtube.com/watch?v=Ddh3p185KhA&t=1667s) **Empowering Casual Coders with AI** - The speaker argues that AI‑assisted programming lowers barriers, creates a rewarding feedback loop for time‑pressed hobbyists, and resonates with anyone who remembers the hassle of legacy coding methods. - [00:30:49](https://www.youtube.com/watch?v=Ddh3p185KhA&t=1849s) **DeepSeek Challenges AI Scaling Law** - The speakers contrast the prevailing belief that larger models and massive compute are essential for performance with DeepSeek’s demonstration that smaller, efficiently engineered models can achieve comparable results using techniques like quantization and distillation. - [00:33:53](https://www.youtube.com/watch?v=Ddh3p185KhA&t=2033s) **Optimizing GPUs Over Raw Compute** - The speakers argue that advancing performance now depends on sophisticated hardware and software optimizations—such as warp specialization in GPUs—rather than merely increasing computational magnitude. - [00:37:03](https://www.youtube.com/watch?v=Ddh3p185KhA&t=2223s) **Quality Over Scale in Model Training** - The speakers argue that careful data selection, refined fine‑tuning, and engineering efficiencies can make smaller, well‑trained models outperform much larger, poorly trained ones. - [00:40:12](https://www.youtube.com/watch?v=Ddh3p185KhA&t=2412s) **Perplexity's Browser Bar Strategy** - The speakers discuss how Perplexity may seize control of mobile browsers' address bars to route searches to its engine, heralding a shift toward voice‑first, AI‑driven user interfaces that could diminish traditional app usage. - [00:43:16](https://www.youtube.com/watch?v=Ddh3p185KhA&t=2596s) **Perplexity AI Replaces Browser** - The speaker explains how Perplexity’s deep OS integration has overtaken conventional browsing for research tasks, but highlights its shortcomings in delivering personalized shopping information that Google currently provides. - [00:46:16](https://www.youtube.com/watch?v=Ddh3p185KhA&t=2776s) **Balancing AI Features and Phone Cost** - The speakers debate how to pack sophisticated AI models into affordable smartphones, stressing democratization, privacy, multimodal interaction, and avoiding past missteps like the Amazon Fire phone. - [00:49:19](https://www.youtube.com/watch?v=Ddh3p185KhA&t=2959s) **Concluding Thoughts on Perplexity** - The hosts wrap up the episode by highlighting market resistance, personal iPhone loyalty, their intent to monitor the Perplexity project, and thanking guests and listeners. ## Full Transcript
0:00Is Manus AI a second DeepSeek moment? 0:02Vyoma Gajjar is an AI 0:04Technical Solutions Architect. 0:05Welcome back to the show. 0:06What do you think? 0:06Almost. 0:07Great. 0:07Kaoutar El Maghraoui is a Principal Research 0:11Scientist and Manager at the AI Hardware Center. 0:14Uh, Kaoutar, welcome back as always. 0:15Uh, Manus AI, what do you think? 0:17I don't think so. 0:18And last but not least is Chris 0:19Hay, Distinguished Engineer and 0:21CTO of Customer Transformation. 0:22Chris, DeepSeek moment, yes or no? 0:25Yes, but no, but yes, but 0:26no, maybe, yes, no, maybe. 0:28Well, we'll be investigating all that 0:30and more on today's Mixture of Experts. 0:37I'm Tim Hwang and welcome to Mixture of Experts. 0:40Each week MoE gathers just the nicest and most 0:42brilliant people to talk through the biggest 0:44news in the world of artificial intelligence. 0:47As always, there's going to be a ton to cover. 0:48We're going to talk about vibe coding, 0:50scaling laws, a new phone from Perplexity. 0:53But first I really want to talk 0:55about Manus AI, which was the focus 0:57of our initial kickoff question. 0:59If you've not been watching the news, Manus 1:01AI is a Chinese company that announced a multi 1:04purpose agent, um, that has really been kind of 1:07taking the sort of AI chatter class by storm. 1:10Uh, they have a bunch of demos showing 1:12their agent able to pull off some 1:13traditionally quite difficult tasks. 1:16So they show it, you know, scheduling trips, 1:19uh, doing stock analysis, reviewing resumes, 1:22evaluating insurance. 1:24And so it's a really, really 1:25wide range of outcomes. 1:27And it seems to be another moment where 1:30following hot on the heels of DeepSeek 1:33is another time when people have been 1:35kind of asking, like, Is China really 1:37catching up, if not surpassing a lot of 1:39the companies that we talk about almost 1:40every single episode on the show at MoE? 1:43Um, you know, your open AIs 1:43and Anthropics of the world. 1:45And so, I guess, Vyoma, maybe I'll start with 1:47you because I think you were the most bullish. 1:49I think you were like, it's close 1:51to being a DeepSeek moment. 1:52Um, do you want to kind of lay out 1:54sort of the bull case here for, for 1:55why it actually is a really big deal? 1:57Um, sure. 1:58So as you know, half of Silicon Valley is 2:01building, um, agentic AI startups right now. 2:04And Manus AI is an agentic 2:07paradigm that we are seeing. 2:08It is more of like an industrialization of 2:12intelligence that has been created from all 2:14these large language models that we are seeing. 2:16If done right, like if they can work well 2:19on the compute side, the hardware side, 2:22they can come up with something because 2:24again, they're first in this entire way, um, 2:27paradigm of bringing it out to the market. 2:29I know like there are so many 2:30other agentic frameworks available. 2:32So I feel that. 2:34If everything goes right in like 10 other 2:36aspects that we have to evaluate from the 2:38metrics, hardware, the software, the compute, 2:40etc. Maybe, but then, as I said, there are 15 2:44others, or 50 others who can always catch up. 2:47So, you never know. 2:48Yeah, definitely. 2:49Um, Kaoutar, you were a little bit more 2:51skeptical, I think, in the opening. 2:53Um, curious about what you think here. 2:55You know, a friend of mine was kind of saying, 2:57it's easy to have a really cool looking demo. 2:59Um, and like a real product 3:01is like a whole nother thing. 3:02And we don't really know whether 3:03or not Manus can deliver. 3:05Is that kind of the source of your skepticism 3:07or is it coming from somewhere else? 3:08Yeah, I think I'm, I'm still 3:10a bit skeptical about this. 3:11I think from my perspective, Manus, um, is 3:14definitely shaking things up here a bit. 3:17I mean, of course there is a lot of 3:18also scepticism in the AI community. 3:20Some argue it's transformative. 3:22pushing the boundaries of 3:23what the AI agents can do. 3:25Others just say it's just a rebranding of 3:27what's maybe, uh, uh, the cloud wrapper, uh, 3:31or cloud is doing more like smoke than fire. 3:35The big question here is can madness 3:37really redefine the AI autonomy, 3:39or is this just another step in an 3:41ongoing AI race between East and West? 3:45So. 3:46Is it just, you know, a leap or it's 3:48just, you know, more advancements here? 3:50Um, so I think there is a lot more evaluation 3:54that needs to be done, uh, to see whether 3:57we're seeing, uh, new innovations, a leap, or 4:00just, uh, kind of maturing up this technology. 4:04So the community is really 4:06interested in the implications for AI 4:08agent development. 4:10So if Manus proves to be a significant 4:12advancement, it could accelerate the creation 4:14of more, um, sophisticated and capable agents. 4:18But of course, there is a lot of pressure here. 4:20So there is growing awareness of the increasing 4:23competition from Chinese AI companies. 4:26Uh, you heard from Vyoma that, you know, half 4:28of the startups are agentic, uh, AI companies. 4:31So there is a lot of competition here. 4:34And I think a lot right now are analyzing the 4:36output of what, uh, Manus is doing to see if 4:39they can see the hallmarks of Claude's outputs. 4:43If this is the case, it's really diminishing 4:45the hype surrounding, you know, this product. 4:47Yeah, for sure. 4:48And I think it's a good 4:49chance to bring Chris in. 4:50I mean, on this kind of Claude point, 4:52uh, the background on all this is that, 4:54pretty soon after Manus came out, people said, 4:56well, there's a bunch of responses here that 4:57are like very kind of Claude flavored, uh, and 5:01in some cases we're actually able to kind of 5:02like pull out sort of like some verification or 5:05some strong evidence that it was from Claude. 5:08Let's say for a moment that 5:09it is just a Claude wrapper. 5:10Does that kind of totally diminish 5:12this as like a, an outcome? 5:14Like I don't know how we 5:15should think about that. 5:16I know a lot of people said, "Oh, if 5:17it's just the wrapper, then, you know, 5:18Manus really hasn't added all that much."" 5:20So I guess from my perspective, I mean... 5:23Let's think about Cursor, for example. 5:25Let's think about Klein. 5:27Let's think about Perplexity. 5:30We could probably argue all of them 5:31are Calude wrappers as well, right? 5:33Which is, you know, they're all tools where 5:36ultimately Claude is driving the experience. 5:39But actually, I don't think this is a 5:41story about which AI model is powering 5:44it, although I think that is important. 5:46This is really a story of somebody bringing 5:48together a really great experience and I think 5:51they have brought together a great experience 5:52because when you use the Manus UI, it does 5:56the planning and it's got a little to do list 5:58and it ticks and it ticks it off as it goes 6:00along and then it's going to have access to 6:02the tools, it'll access the terminal, access 6:04the browser, very similar to what's going on 6:06with, you know, OpenAI with Operator, etc. 6:10and Deep Research, for example. 6:12They brought that together in a nice experience. 6:14They're running it on a sandbox. 6:16They're doing tool calling 6:18and it kind of feels good. 6:19Right. 6:20And now it's a little bit more than 6:22a Claude wrapper to be fair to them. 6:23They have, you know, taken the 6:25open source tools and they've 6:27integrated them really well together. 6:29Right. 6:29So, um, so technically we 6:32could go and do this ourselves. 6:34And I think that's. 6:35Why this is probably gonna 6:37end up, this is why I went. 6:38Yes. Maybe, maybe, maybe yes. 6:39Yes. Maybe. 6:40'cause I think what's gonna happen is the open 6:42source community is gonna go, we can do that. 6:44And, and I know that 'cause I've been coding 6:46away all week, uh, doing the same thing. 6:48Right. 6:49Trying to do the same thing. 6:51Yeah, exactly right. 6:51Like, like every other developer on the planet. 6:54Right. 6:55And, and therefore. 6:56It's something that's achievable. 6:58And, and to be fair to them, 7:00it's a little bit more as well. 7:01They said that Claude was doing 7:02orchestration, but they also said they 7:03fine tuned a bunch of Qwen models. 7:05And I think specifically said that for the, 7:08the planning model, the, the one that sort of 7:11comes up with the to dos, et cetera, that 7:13was a particular kind of Qwen fine tune and 7:15they pointed to a version they'd done earlier. 7:18So, so it's a little bit more than 7:20just, here's Claude with a pretty UI. 7:22It's a bunch of fine tuned models. 7:24It's, you know, bringing these tools 7:25together, sandboxing it, and then 7:27bringing the package in together. 7:28And I think they've done 7:29a kind of a fabulous job. 7:30And then finally, they've 7:32generated the hype, right? 7:33I mean, I was reading today, it had like 7:352 million folks have signed up for invites. 7:38So, we're all running going, 7:40yeah, we're going to do this. 7:42Will they be around? 7:44Are they the next Devon? 7:46You know, we will find out in six months or so. 7:48Right. 7:49But, but for just now the hype 7:50cycle's there, but I'm hoping it 7:52galvanizes that open source community. 7:54And you had a good phrase there, which 7:55is, you know, they're tying together 7:56a bunch of components and like, 7:58sort of we could do this as well. 8:00Um, and I did want to dive a little bit 8:02into that because I have a friend of mine 8:04who you know, he focuses a lot on all the 8:06kind of like state level AI bills that 8:09are kind of bubbling up. And he made the 8:11observation where he's like, well, look we 8:13the US Companies could have done this US 8:15open source kind of efforts could have 8:17launched a very similar thing you know, 8:20maybe one reason they don't is because like 8:22some of the things that Manus is showing 8:23off, you know have been kind of risky from 8:26a legal standpoint in the US, right? 8:28Like things like resume review is 8:30like a thing that, you know, kind 8:31of is like very hotly regulated. 8:33It's a hotly disputed thing. 8:35And so I guess maybe I'm 8:36going to toss it back to you. 8:37Like, do you think almost there's 8:38kind of like an edge here? 8:39Like almost like Manus is winning 8:41this or in the very least they kind of 8:43seem like first to getting this like 8:44hype wave just because they've been 8:46willing to be more aggressive than other 8:48folks or do you do not really buy that? 8:50I feel the first thing that Anthropic 8:52cloud had tried something called computer 8:54use we spoke about it in this board with 8:56you and which is sort of compared quite 9:00vigorously right now with Manus AI but 9:04the computer use Anthropic cloud version it 9:06actually performs very well in controlled 9:08environments exactly what we're talking about. 9:11Like, let's say if there's resume review, 9:12et cetera, it brings along a whole different 9:15metric system that has to be evaluated 9:18for a large language model to be used. 9:20Like a use case, POC is very different 9:24from what can you integrate in an 9:26enterprise architecture, right? 9:27And how do we integrate it? 9:29So sure, Manus showed the way that, okay, yes, 9:32this is how we can do it, but to actually do it. 9:35It's going to be a lot of leaps and bounds 9:38that the entire industry has to go through, 9:40regulations, et cetera, uh, have to be 9:42written around it for us to be able to use it. 9:45But yes, the US did try it and that the 9:48computer use part was actually something that 9:50we were all talking about for a while, right? 9:52So Kaoutar, I guess maybe the final kind of 9:55question and curious to get your thoughts 9:56on this is whether or not like, take us six 9:59months into the future, like Chris is saying. 10:01Like, do you have predictions 10:02on where this all kind of goes? 10:04I mean, one thing's for certain. 10:05It seems like we're going to see a bunch of open 10:06source attempts to kind of do the same thing. 10:09I guess we can ask the question of whether 10:10or not this Manus thing actually changes the 10:12fundamental trajectory of where agents is going. 10:15Um, but kind of curious if you want 10:16to paint a picture of like, where 10:17you think we will be in six months. 10:19You know, if anything, Manus is kind of just 10:21building more hype and more momentum in this 10:22direction, but curious to get your prediction. 10:25Yeah, I think it's a good question. 10:26I definitely feel, you know, I agree with Chris. 10:29That's, you know, what they're doing is, 10:31it's not just integration, but it's also 10:34more about having this fully autonomous. 10:36agent capable of independently executing 10:39complex tasks, doing various things 10:41like sorting, stock trend analysis, 10:43website creation, which is really great. 10:46So we will see others trying to mimic that. 10:49Uh, and they're leading in this space 10:53around, you know, this autonomy. 10:55Uh, so beyond mere integration, making a 10:58significant advancement in the AI autonomy. 11:01But I think whether, uh, more 11:03hype will follow, I think. 11:06It's gonna be the case. 11:07Uh, we're seeing now every few days or 11:11every few weeks we're seeing new hypes. 11:13So I feel we will see, uh, more interesting 11:17things coming in this space here. 11:18Yeah, it's like if this is a DeepSeek 11:20moment, then get ready for like at least 11:2120 or 30 more this year, I suppose. 11:24I just want to say, you know what 11:25I don't want to see in six months? 11:27Another browser operator. 11:29You know what? 11:30Large language models are 11:31really good at text, right? 11:33Why? 11:34Why are we insisting we have an AI moving a 11:37cursor around, finding a bounded box, taking 11:40a screenshot, and then typing into the box? 11:43You know what I would rather see? 11:44I would rather see somebody go, you 11:46know what, I'm going to create an AI 11:47native browser which parses the text. 11:50And actually, yes, it's going to communicate 11:53with, with, you know, the websites, and they'll 11:54recognize it as a real browser, et cetera. 11:56But you don't need to do screenshots. 11:59You don't need to move a cursor around. 12:00You're a browser and you're a large 12:02language model that knows how to code. 12:03Do that. 12:04That's what I want to see 12:05in the next six months. 12:06People have gotten lazy, Chris. 12:07People don't want to do that. 12:09They're like, if we can have 12:10someone do this for us, why not? 12:12I'll sit on the couch all day. 12:15I'm fine for them to sit on the couch. 12:18Just, just don't move a cursor around. 12:20That's what I'm, that's 12:21what's, and take screenshots. 12:23That's what's bothering me. 12:24Yeah I think that is one of my 12:25favorite agentic tropes at the moment. 12:27I mean people do it because 12:28it looks really cool. 12:29Like that's the main reason is that it's 12:31like it's it's kind of cool and spooky. 12:33Yeah I think it's more for demo 12:35purposes. 12:35And also for people to like show that this is how we can do AGI. 12:40Like this is the next step to AGI. 12:41That's, and I think everyone's chasing that now 12:44that, okay, we're done with the LLMs, et cetera. 12:46Now let's move on. 12:46But Vyoma, it sounds like you're defining 12:48AGI as a 95 year old grandparent trying 12:53to work the internet for the first time. 12:55That, that's what I see when I 12:56see the AI operate with a browser. 12:58Yeah, 12:59that's true too. 13:00I mean. 13:01That's what we defined, I guess, at 13:02this point, but let's break that. 13:04Yeah, I think it's going to be interesting to 13:05see how these human interfaces will evolve. 13:08So I agree with you. 13:10I also don't like the cursor on these things. 13:12So probably more serious thinking 13:15into what would be interesting 13:16for us to see as these interfaces. 13:19What would you like to see? 13:20So it really mimics a true human 13:22experience without having this cursory 13:25or, uh, screenshots and things like that. 13:28So, 13:28yeah, for sure. 13:29Uh, Chris, if for your weekend experimentation, 13:31have you been able to replicate Manus? 13:33I am surprisingly far. 13:35Actually, I went 13:37slightly different from them. 13:38So I've put MCP at the heart of what I've 13:41been doing, uh, which I think is, is a 13:44lot better, but then I haven't built a 13:46product and got it to 2 million people. 13:48This is just Chris and his agents in the night. 13:50So I, you know, I don't think I can take 13:52the intellectual high ground on this, but I, 13:54yeah, I've got, I've got pretty far so far. 13:55Yeah, that's actually, I mean, it's a 13:57pretty, uh, strong indication, right? 13:58Like with not a whole lot of work, you 14:00actually get pretty far with these things. 14:02Um, I guess it just goes to show how 14:03competitive this space is about to be. 14:10I'm going to move us on to our next 14:12topic, which is a real fun one. 14:14Um, Andrej Karpathy, who we've talked 14:15about many times on this show before, 14:18in addition to being, you know, former 14:20Tesla and former OpenAI, I think he's had 14:22this kind of like career job in sort of 14:26shaping the memes of the, uh, AI space, um, 14:29we mentioned Cursor earlier, arguably his 14:32shout out of Cursor is one of the reasons 14:33that Cursor has been so wildly successful. 14:36And, um, he had a nice tweet kind of capturing 14:39sort of his thoughts on kind of using AI 14:42assisted coding recently, where he said: 14:44"there's a new kind of coding I 14:46call vibe coding, where you fully give 14:49in to the vibes, embrace exponentials, 14:51and forget that the code even exists. 14:53It's possible because the LLMs, 14:55e.g. 14:56Cursor Composer with Sonnet, 14:58are getting too good". 14:59Um, and this has kind of been a funny 15:00thing because in, you know, true Andrej 15:03form or Karpathy form, um, the vibe coding 15:07has just kind of like gotten everywhere. 15:08It's just been like a joke that 15:09people keep mentioning now. 15:10Um, and, uh, and weirdly, I feel like in the 15:13last week or so, or however long ago, um, 15:17people have been like, oh yeah, I'm like, 15:18I did this project through vibe coding. 15:20You know, it's almost kind of now 15:21becoming like a, a term of art. 15:23If I may be able to turn it to you first is 15:25like, is vibe coding a real thing? 15:27Like, is this the future of coding? 15:28Is people just kind of like, you know, just 15:30kind of vibing with it until an app comes out? 15:33Um, is this a good way of kind of 15:34thinking about where engineering 15:35is going with all this assistance? 15:37Yeah, it's going to be very controversial 15:39when I say this, but no, it's not. 15:41I don't think this is the future. 15:43And I feel that getting to know the concepts 15:46and the basics behind how a particular code 15:49works is something which is extremely important. 15:53Sure, you you won't be able 15:54to like code it end to end. 15:55You can use wipe coding to assist you with that. 15:58But that being the only uh, crutch 16:01that we all rely on, I don't think 16:03this is going to be successful. 16:05Like sure, I can do wipe coding 16:07for a weekend project to test out 16:09something that I want to maybe show a 16:11small POC to understand the concepts. 16:13But for a diff, and a diff which is like going 16:16through millions of lines of code over there, 16:18you use that particular code to solve that diff. 16:21And then, put it in production. 16:23So totally against that. 16:25I don't think that's the right norm. 16:27The other thing is, I know everyone keeps 16:29talking about white recording, but if you go 16:31back in the interview market right now, you 16:34have to go through a lead code interview. 16:36You have to do solution design. 16:38So the basics aren't going anywhere. 16:40People are talking about it, but yes, you 16:43have to traverse a string if needed or like. 16:45add the nodes of a binary tree. 16:47Yes, you have to do that. 16:48It's not because people want to know 16:50how well you code, but they want to 16:52know whether you understand the concept. 16:55Right? So that's something that I feel, um, is needed. 16:58Yeah, so there's a lot to unpack there. 17:00I mean, I think the first thing maybe 17:02to get into is, yeah, I don't think even 17:04Karpathy is like, oh, you should vibe code. 17:06Vibe code like an enormously complex project. 17:10But I think it is kind of 17:11this interesting debate. 17:11I mean, you sort of say like, look, 17:13this is going to work for your weekend 17:14project, but probably not much further. 17:16I think there's almost a question of 17:17like, where do you draw that line? 17:19Like, how far can you get with vibe coding? 17:22If everybody agrees, you can't do the 17:23most complex thing, but you can definitely 17:25vote vibe code the simplest thing. 17:27You know, this dividing line gets very fuzzy as 17:29these models get better and better and I think 17:31it is, it is a genuinely interesting question. 17:33I'm curious if, are you a vibe coder? 17:35Are you, do you vibe code? 17:37I do it sometimes and I like it. 17:38Actually, I'm really fascinated by vibe coding. 17:41So, um, and I see vibe coding as kind of 17:45a reflection also of the changing nature 17:47of software development where AI tools 17:50are increasingly being adopted and they're 17:52increasingly handling routine tasks. 17:55Uh, allowing, you know, the coders to focus 17:57on higher level design and problem solving. 17:59Of course, I think I'll see we're evolving 18:02into a world where we're going to be 18:04combining both vibe coding and serious coding. 18:07But of course, understanding what 18:09this, what the AI is generating, how 18:11to test it, how to integrate it, how 18:14to do all these, these diffs and so on. 18:16It's still going to be very important. 18:18Whether we're going to get to the point 18:19where it's all going to be automated 18:21by AI, I think it remains to be seen. 18:23Uh, I think we're heading into the 18:25direction as these, um, AI, uh, LLMs are 18:29becoming better and better and coding. 18:31Uh, one of the things that I'm a 18:33bit concerned about is the rigor. 18:36Um, so the vibe coding, you know. 18:38could also lead to this decline 18:40in code rigor and best practices. 18:43So there is a worry, you know, that I see that 18:47are we getting into also less experienced coding 18:50coders that rely more and more on vibe coding, 18:53especially among students, people that are 18:55still learning, uh, you know, they're, they're 18:57given an, uh, programming assignments and 18:59they'll just go and ask, you know, an AI agent 19:02or a LLM to give me the assignments 19:05and then most of these assignments sometimes, 19:07you know, the AI does pretty good job. 19:10So, um, so they're definitely going 19:14to be an influence in these, uh, vibe 19:16coding or AI tools, uh, influencing 19:18the coding styles and the practices. 19:21Uh, but this is also 19:24enabling more exploratory and 19:25iterative approach to coding. 19:28Yeah, for sure. 19:29My friend actually, he, 19:30uh, he coded up this app. 19:31And, uh, I was like, oh, so like, 19:33what, what does this menu do? 19:35And it was very funny because he 19:36was like, I just vibe coded it. 19:38So I'm actually not really sure what it does. 19:40And I was like, I don't know if this is 19:42like a sustainable way to go about building 19:44bigger systems, but this bleeds pretty well, 19:47I think, to Vyoma, a point that 19:48you made, which is first you said, 19:50okay, it's only going to be good for 19:51weekend projects, not anything bigger. 19:54I think the second point that you made 19:55is also really interesting as well. 19:57If you want a job as an engineer, they're still 19:59going to make you go through LeetCode, right? 20:01Like you still are forced to kind 20:02of like go through this gate. 20:04And I was joking with a friend recently. 20:05I was like "Oh, well, I'm just 20:07looking for the 10x vibe coder". 20:09Right? 20:09Like someone out there is able to vibe code 20:11like way more proficiently than everybody else. 20:14If I can just find that person, 20:16um, maybe he doesn't, he or she 20:17doesn't need to know LeetCode. 20:19So even, I'm not saying that LeetCode 20:21is 100% a reflection of how 20:24good, uh, a software engineer you are. 20:27But even if you're not able to solve that 20:29pseudocode, how does that if a statement work? 20:32How does this while statement work? 20:35Trying to explain what you are asked from a 20:38LeetCode question is also good enough in 20:40case you're not able to code on spot during 20:43that particular time that I'm asking you. 20:45A 10x viber, vibing coder that 20:49we are naming it, I don't know. 20:51If that particular coder would actually 20:53even understand a legacy code that has 20:56been written because for that you'll have 20:57to go back and understand, okay, what's a 20:59function, how did that function, what are 21:01the parameters called in this function, 21:03how is a parameter written here, so I'm not 21:06there yet, maybe I have not seen a good use 21:10of that entire, uh, system, but 21:12maybe, um, people might get better. 21:14The models keep getting better. 21:16I totally agree with that. 21:17Sure, you can use it for grunt work, like if 21:19that was one of the questions that, uh, was 21:21coming up when I was reading about it is like 21:24if there is a CSS file and you need to change, 21:26uh, a particular bracket or a button, you 21:29don't have to sift through thousands of lines. 21:30Of course, sure, you can use it for that 21:32because obviously better use of your time to 21:34do something else, learn something better, but 21:37for you to build an application end to end, 21:40I don't know if it's the best way to do it. 21:41I can say right now, this 21:43is, this is here to stay. 21:44This is not a weekend project thing. 21:46And, and as somebody who's starting 21:49graduate job was writing Unix C motif, and 21:53I, and I can tell you right now, I spent 21:56most of my time chasing memory pointer 21:59bugs, right, you know, and I am never 22:03going back to that nightmare ever again. 22:06I, you know, you talk about productivity. 22:09You look at your own terrible graduate 22:11written code and try and work out 22:14why that memory location is not 22:16the place you wanted to point at. 22:18You know what, I, and you watch your dreams 22:21die as every time you boot up your application 22:24and the thing goes, wow, wow, wow, wow. 22:28You know, and what, what. 22:29What I can do with a large language model 22:31is I can go Ctrl+C, Ctrl+V, and with 22:37the error message from the compiler, and 22:39then say, fix this buddy, and then it goes, 22:42ah ha ha, you, you messed it up over here. 22:45Oh, that's great. 22:46I will just copy and paste that back in. 22:48Or using Cursor, I don't even need 22:50to copy and paste, I can just go. 22:51Click, click, click, like Homer 22:53Simpson in the nuclear power factory 22:55when he was working at home with that 22:58little pigeon going dun, dun, dun. 23:00That's, that's the world we're moving to, right? 23:04So I'm, I'm all in. 23:06I'm all in. 23:08Okay, there's about, you've 23:09got some inbound here. 23:10Uh, Kaoutar, how about you go first? 23:12I'm a bit worried about also if we only 23:15do this, then if you really want to do 23:17software hardware co design, really doing 23:20optimizations, then people will lose the 23:23skills to understand the computer architecture. 23:25What does it mean to have a dangling pointer? 23:28How do we do these memory 23:29allocations, the efficiency? 23:31I mean, I see that right 23:32now as we are designing. 23:34these efficient systems, having a core 23:37understanding of these concepts is 23:39very critical to know how to optimize. 23:41So if we're just going to do vibe code 23:43and then people have no clue what, you 23:46know, underlying systems are doing, how 23:47they're behaving, what does it mean to have 23:50a cache hierarchy, you know, all these 23:52implications about, you know, the data movements 23:55and the computational units, the bottlenecks. 23:58We're going to lose that. 23:59And that's what worries me. 24:00And how do we optimize these systems end to end? 24:03I'm okay with losing that, to be honest. 24:05I'm still having, 24:07yeah... 24:07Of course, it's not for everyone. 24:09I'm having nightmares from, from my past. 24:11there. You know, so no, I'm fine with that. 24:14But then, but honestly, I think, I 24:15think those skills are important. 24:17I think we have different, uh, we have different 24:19skills within the engineering practice, 24:21and I think that's going to expand out. 24:23And, and if you really think about it, I 24:25mean, like, if you think of something like 24:28Lewis Hamilton as a Formula One driver, 24:30one of the best drivers in the world, etc. 24:32Does he know how to change a tire? 24:34I don't know. 24:35Maybe he does, maybe he doesn't, right? 24:36But, but what he does know how to do is drive 24:39a Formula One car at speed through, you know, 24:42uh, the race course and better than anyone else. 24:45Now, the question I would have there 24:47is that there is a wider team, right? 24:49Some people are going to be 24:50specialists at tire changes, some, 24:51you know, ball bearings, whatever. 24:53And I think that's, that's pretty cool. 24:55So, not everybody needs to know how 24:58to deal with dangling pointers, right? 25:00Some people just want to build an app and 25:02get it out and try and make some money. 25:03Go for it. 25:04Yeah, there's some interesting history 25:05here too, because it's almost like, I 25:07guess, you know, just think about like 25:09the first people who built planes, right? 25:10Like the Wright Brothers. 25:12They were like, you know, engineers, right? 25:13And they were like, you know, modifying 25:15bikes to get the plane to work. 25:17And like, yeah, if you're flying a plane, 25:18then it almost breaks down so often that you 25:20really need to understand every bit of it. 25:22Right. 25:22And then now pilots have a certain 25:24level of training, but they're not 25:25necessarily airplane engineers. 25:27I actually wonder if that's going to kind 25:28of interestingly have that divergence in 25:30coding over time, where you almost have 25:32like good coders, but that's almost like a 25:35discipline that's almost separate from like 25:37understanding the inner workings of the machine. 25:39And. 25:40You know, I guess we get that in part because 25:42like machines become super reliable, right? 25:44Like your car is not breaking 25:44down every single day. 25:46So you don't necessarily need to understand. 25:47But you know, software used 25:48to break down all the time. 25:50And so you really do need to 25:51know those internal components. 25:52And first of all, I'm very happy to know 25:53that Chris is a fellow F1 supporter. 25:55So even if Hamilton doesn't try to fix a car, 26:00Chris, I think he would know how to do it. 26:02Like theoretical concept. 26:03That's all I feel is. 26:04actually needed in the wipe coding part and 26:07the other thing is I wouldn't entrust anyone 26:09with like a wipe coder with the nuclear 26:12reactor control system, but I would make 26:15sure that, um, anyone who does wipe coding 26:18actually knows what they are doing with it. 26:21So as long as that loop ties 26:23back, I, we are okay with it. 26:25But if it doesn't, and we have someone who's 26:27just like vibing with in minimum amount of data 26:32points and trying to make something production 26:34ready, I don't think I'm comfortable with that. 26:36I, I think, I sort of agree with you 26:38and I think, I think we're going to 26:39move from one extreme to the other 26:41and I think that's the reality, right? 26:43And I, I can see a world where you're going to 26:46vibe code to prototype, you're going to vibe 26:49code to figure out some issues, etc. And then 26:52it's almost, I think, I've never painted, but 26:54it's like, you know, those like Monet paintings 26:57or whatever, and then it's all sort of blurry 26:59stuff, and then you sort of hone into the 27:01detail, I think that's going to start to become 27:04a bit of a pattern, right, which is like, okay, 27:07I kind of need something like this, I'm going 27:08to do this, you're going to orchestrate it. 27:10And then you're going to start to say, 27:12okay, I know what I've, I've built here. 27:14I prototype this. 27:15Now I'm going to start engineering this 27:16further and then go down and detail. 27:18And so I think there's probably a hybrid model, 27:21but then the flip of this is, is we're 27:23looking at this from an engineering perspective. 27:26Why can't that person who's never coded 27:30not go and create an app for themselves and 27:32get it in the app store and make some money or 27:35maybe somebody who wants to do a home automation 27:38project, but has never had those skills. 27:40Why can't they vibe code and then 27:43be able to do that thing that 27:45they've never been able to achieve? 27:47And then you know what? 27:48It might get them interested in the discipline 27:50and might want to say, you know what? 27:53How does memory work, right? 27:55And then they start delving into that. 27:56So I, I think actually it's an area 27:59where we can have a greater inclusion and 28:03greater impact and, and a larger community. 28:06So I, I, I hope that's the direction we go in. 28:09Yeah, I think the feedback 28:10loop is super important. 28:11Um, I mean, at least for me who like doesn't 28:13really code day in and day out, like the 28:15ability to use these tools just makes the 28:17experience of like, I have 45 minutes after 28:19my kids have gone on down to like kind of play 28:21with the computer and it's like oh I could 28:23just like get further in that time and it's 28:25just like a very strong feedback was like 28:27very satisfying in a way that kind of like pre 28:30these tools like you didn't really have. A final 28:34anecdote that y'all might find interesting. 28:36So my mom was like a very early coder. 28:39And so she still remembers like, you know 28:40punch cards and it's it's Chris to your 28:43point It's like it's interesting to me how 28:44much like if you felt a lot of pain coding, 28:47you are more likely to want these tools 28:49because you remember how painful it is. 28:51Like my mom's response is oh, yeah I remember 28:53programming a big box, like, of punch cards, 28:56dropping them, and then basically spending, 28:58like, hours having to, like, recompile. 29:00And she's like, I love this. 29:01Like, we just, like, automate 29:02everything, basically. 29:03Um, which I think is, like, kind of a 29:04really fascinating aspect of, like, people's 29:06personal experience of just, like, how 29:08difficult it is, might make them more 29:10or less willing to adopt these tools. 29:12And you could imagine a 29:13vision model, multimodal, controlling an 29:17action model with a robotic arm, and then 29:20that robotic arm can resort those punch 29:22cards, and your mom would be fine, right? 29:24And that's right. 29:25Yeah, exactly. 29:26We should. That's right. 29:28I need, I need the, the coding 29:29assistant, but for, for punch cards, that 29:31actually would be an awesome project. 29:32I agree with Chris that it's going to be a 29:34hybrid world where we have these people who 29:37have no clue about coding, but they're still 29:39being able to use this vibe coding to create 29:41really nice things for rapid prototyping 29:43or proof of concepts or applications maybe 29:45that have been mature by these, uh, tools. 29:49But then we still need the people who 29:51really understand what's, what's happening 29:53behind the scenes, who can know how 29:55to debug, who can know to optimize. 29:57It's going to be specialization 29:59at these different levels. 30:00For sure. 30:00Yeah. And I, I got to believe, I mean, some of 30:02these debates happened when like object 30:03oriented programming came around, right? 30:05People were like, ah, you don't 30:06understand the DNR workings of the system. 30:08It's like this battle happens almost at 30:09each layer of abstraction, uh, arguably. 30:16I'm going to move us on to our next topic. 30:18Uh, we wanted to do a quick segment to 30:20talk a little bit about scaling laws 30:22and this segment kind of puts together. 30:24I think a couple of things that 30:25we've touched on last few episodes, 30:28particularly when it comes to DeepSeek. 30:30Um, and I think Kaoutar, great to have you on the 30:32show because I think you suggested this topic. 30:34Um, the background here of course is, uh, 30:36scaling laws are really the idea that we 30:39have kind of this interesting relationship 30:41in machine learning where sort of the, the 30:43more compute that you're using in doing 30:45pre-training the better the 30:48capabilities are that kind of come out. 30:49There's kind of this rough relationship 30:50between like how much kind of like 30:52muscle you're putting in and the model that kind 30:55of comes out of it. And you know, this has been 30:59kind of like the thing that has motivated the 31:00entire thesis of these companies, particularly 31:03the kind of frontier model companies raising 31:05enormous amounts of money, which is to say 31:07well if we want really powerful systems, we 31:09really need lots and lots of data lots and 31:11lots of compute and we need to do the biggest 31:13possible sort of pre-training run that we 31:15need to do and I know Kaoutar, you wanted 31:17to bring up this topic because you want to 31:19talk a little bit about how you think that 31:22DeepSeek kind of doesn't necessarily break this 31:25idea, but kind of nuances it a little bit. 31:27Do you want to talk a little bit about that? 31:28Yeah, definitely. 31:29So of course, as you mentioned, in traditional 31:31AI development, there is this general belief 31:34that bigger models and larger data sets lead to 31:37better performance following the scaling laws. 31:39And this often translates into these massive 31:41investments in hardware and infrastructure. 31:44But what this DeepSeek really 31:45demonstrated, uh, they're challenging 31:48this traditional AI scaling laws. 31:51They're demonstrating that smaller, 31:53more cost efficient models, they 31:56can achieve competitive performance. 31:58And they're, you know, even this titan 32:01existing business models, reliance 32:03on these large scale infrastructures. 32:05So they, they used a lot of techniques, uh, 32:08like, uh, uh, quantization and distillation 32:11and, uh, they even, you know, did some 32:14optimizations at the PTX level, given the 32:17limitations that they had with the H100 GPUs. 32:20So a lot of focus on efficiency, 32:22oversize, so, uh, emphasizing efficiency 32:26rather than the sheer model size 32:28and optimizing at different levels in the stack 32:33and also looking at, you know, 32:35enhancing the data quality and 32:37employing better training strategies. 32:39So I mean, the, the, the key idea here is how do 32:43we leverage smarter techniques, uh, or smarter 32:48training techniques, for example, for their, in 32:50their example, that achieves better performance 32:53with fewer parameters and reduce these 32:56computational costs and computational needs. 33:00And I was really fascinated by the wide 33:02range of techniques that they have used. 33:04You know, they had these 33:05data centric approaches. 33:07They also had the hardware wear optimizations. 33:10They were considering also sustainability. 33:12Um, and I think this has, you know, 33:15implications on the AI community. 33:17So where the focus here is shifting from the 33:20scaling by size to scaling by efficiency. 33:23I think this is actually, I mean, like 33:25there's a lot here that we could get into. 33:27Um, I think in some ways for me, the kind of 33:30scaling law question is kind of interesting 33:32because it has kind of turned out that 33:34people have meant a couple of things when 33:35they say scaling laws, um, in, in the 33:38popular, uh, kind of discussion of AI. 33:41And, 33:41you know, one of them is just 33:43like how much compute you need. 33:45Um, and I guess in some ways 33:46DeepSeek doesn't really change that. 33:48It certainly changes the kind of platforms you 33:50need to get high performance out of the models. 33:54Um, but I guess it doesn't really 33:55necessarily kind of eliminate 33:57the idea that like more compute, 33:59like, equals better performance. 34:01Is that, is that right? 34:01Is that the right way of thinking about it? 34:03I don't know, Vyoma, if you want to jump in. 34:04With the scaling laws, with DeepSeq, 34:06with etc. I see a new shift in 34:09people trying to optimize GPUs. 34:12Or, uh, the ways in which they can 34:14revolutionize this entire field. 34:16So I don't know if, uh, people know about 34:18this, but a month ago, Meta and NVIDIA 34:21came up with a paper and with some, 34:23and they said something called as warp 34:25specialization will be a part of PyTorch. 34:28So it kind of optimizes the GPU performance on 34:31the, any of the hopper architectures, like the H100s that they do by assigning like some 34:36sort of distinct role to each one of these warps. 34:40And what is one warp? Like a group 34:41of 32 threads that are running. 34:44So it kind of pivots to this entire point that 34:46we are looking into how do we optimize 34:49all of these hardware specs, which are 34:51also available based on the 34:54previous information that we've got. 34:55And I think that had it also came 34:57into picture because of some of the 34:58scaling loss that we've been seeing. 35:01So I don't see that as a way in 35:03which it would limit us, etc. I 35:05feel we've come up with better ways. 35:07Right. 35:09It's not necessarily about magnitude. 35:11It's more like how we're 35:11treating the GPUs basically. 35:13Exactly. 35:14Exactly. I think the hardware where optimizations 35:16are becoming increasingly important. 35:18If you also see the work that's 35:20happening around these state space 35:22models and, uh, the flex attentions. 35:25Every now and then we hear about, you 35:27know, different algorithms around flash. 35:29You know, how do you do these transformer 35:31attention computations more efficiently? 35:33Like there is a flash attention. 35:35There are various versions 35:36of these flash attentions. 35:37There is a flex attention. 35:38Now the Mamba and the Bamba models. 35:42They're also doing a lot of optimizations by 35:45understanding the underlying architecture, 35:48especially the GPUs right now, and then figuring 35:51out how to restructure the computations, 35:54and especially the data movement so you can 35:56drive more efficiency from the hardware. 35:58Um, so also other things 36:02like the test time computes. 36:04which is something that is 36:05also becoming very important. 36:06So instead of focusing solo 36:08solely on pre-training computes. 36:10So, and this is an example also that DeepSeek, 36:13uh, uh, pointed out, which can we focus on 36:18inference time computes, which is really 36:20more critical, meaning smaller models can 36:23compute more at test time, longer reasoning. 36:27Tree search, Monte Carlo 36:28inference, and things like that. 36:30And this also reduces the need 36:31for enormous parameter count. 36:33There was a very interesting paper about 36:35test time computes, which showed all of 36:37these techniques that really focuses on how 36:40do we bring more from the model during test 36:42time, not during the pre the training time. 36:45And also the distillation, this creating these 36:48compact models with large model capabilities. 36:51So, of course, you still need to have the 36:53large model, but we can create a variety 36:55of distilled versions that do really 36:58better and can also inherit knowledge 37:01and reasoning from much larger models. 37:04And of course, you know, I think the high 37:05quality training is also something that 37:08is outperforming this raw scaling, smart 37:11data selections, better fine tuning, 37:14reinforcement learning with self improvement. 37:16So well trained models can also outperform 37:19these poorly trained massive models, 37:21especially if you focus on the data quality. 37:23Yeah, I was just going to say I think actually 37:25we've came from a world of vibe training, 37:28which is really, if you think about what 37:31was going on in the beginning, which is just 37:33like, we're going to take some transformers, 37:36and we're just going to throw a bunch of 37:37data and get it to NextSoak and predict. 37:39And actually we're in this stage now 37:41where it's really about honing the 37:43algorithms, honing the chain of thoughts, 37:44as you say, starting to engineer things. 37:46I mean, Kaoutar, you made some really 37:48good points on DeepSeek, right? 37:50So actually one of the interesting things 37:53they did, I think it was last week, is 37:55they open sourced a whole bunch of their, 37:57uh, code bases that you use to train. 37:59So, and that's everything from data frameworks, 38:02they, they even engineered themselves a new 38:04file system, a distributed file system, etc. So, 38:07actually, all of these engineering techniques, 38:10anything you can get more efficiency, anything 38:12on a better training, I loved your point about 38:14the high quality chain of thoughts and inference 38:16time compute, that makes a huge difference. 38:18That allows you to, to then. 38:20start getting smaller models, higher 38:22quality models, and I really think we've 38:25moved into this kind of engineering phase. 38:27So, but I'm going to, I'm going to, 38:28like Karpathy, I'm going to, I'm going 38:30to call vibe training and see if, see 38:32if I can, see if I can get myself a 38:34Wikipedia entry off the back of that. 38:36Yeah, you heard it here first. 38:37Um, I guess Kaoutar, maybe I'll throw it 38:39to you for the last kind of question here. 38:41Do we think that scaling laws no longer matter? 38:44Like, do we care about scaling laws anymore? 38:46Given this kind of new era of 38:47optimization that we're now in? 38:48Yeah, I think we should really 38:50shift from just bigger to more 38:52smarter and more efficient models. 38:55So I, we should really redefine the 38:57traditional scaling laws as they were 38:59defined by just bigger, better, but 39:02I think it should be about 39:04smarter and more efficient. 39:11Well, I'm gonna move us on to our final topic. 39:13I want to end on sort of a, a kind of 39:15fun, sort of odd story that kind of 39:17came across our um, uh, sort of cues. 39:20Um. 39:21There was an announcement recently that, 39:22uh, Deutsche Telekom and Perplexity, um, 39:26were going to work together to announce 39:28and launch, uh, what they call an AI phone, 39:31which would be a phone that integrates 39:33a bunch of, I guess, AI features for 39:35less than 1,000 and coming out in 2026. 39:39Um, and this news kind of struck everybody as 39:41like a little bit sort of surprising in some 39:43ways, because if you know Perplexity, the 39:44company, um, they largely have been in the world 39:47of, uh, AI search, AI powered search, right? 39:50I think they're one of the first to market 39:51in terms of, you know, you ask a human 39:53language query and it gives you sort of 39:55results, um, that, you know, attempted 39:58to kind of be better than what you'd get 39:59from the sort of quote unquote sort of 10 40:01blue links or 10 links from, from Google. 40:04Um, and so I think the first question, 40:06which is that it's a very kind of 40:07perplexing announcement for Perplexity, 40:10uh, to be getting into the phone space. 40:12Um, Chris, you're already smiling, 40:14so maybe I'll throw it to you first. 40:15Why is Perplexity doing this at all? 40:18Tim, when you're on your mobile 40:20phone and you're doing your Google 40:21searching, do you ever go to 40:25www.google.com 40:26and then type in your query there? 40:29Is that your action? 40:31I do never, 40:32I never do that. 40:33What, what is your action, Tim? 40:35How do you 40:35search on a regular mobile phone? 40:38Uh, I would say open the browser and then 40:40I type my search term into the browser bar. 40:43Exactly, so this is what this is really 40:45about is controlling the browser bar, right? 40:48So, at the end of the day whoever 40:51controls the browser bar means 40:53that they can direct those queries 40:55to their search engine. 40:57So, if I was Perplexity, I would 41:00absolutely launch a mobile phone where 41:02I'm in control of the browser bar. 41:04That's, that's my opinion of what 41:06they're doing, and that is a smart 41:08move. 41:08Uh, do you all agree? 41:09Vyoma, Kaoutar? 41:10Uh, curious if you are like, 41:12brilliant move by Perplexity. 41:13This is exactly what should happen. 41:16And I think this is also marking a shift 41:18in the user interface, you know, how these, 41:20how we're interacting with our phones to 41:24shift to a more voice centric AI driven 41:27user experience, potentially reducing also 41:30the reliance on the app based interactions, 41:33like, you know, going to apps and so on. 41:36I think maybe this is going to shift 41:37and change with the, these AI phones. 41:41It's going to be maybe a completely different 41:43experience that is mostly voice centric. 41:45And probably we'll see the 41:47disappearance of the apps. 41:48And more of these agents in the background 41:51working together to satisfy whatever we need. 41:54Yeah, the interface part of this I 41:55think is really super, super interesting 41:58and it's something I actually want 41:59to dig into a little bit more. 42:01You know, one thing that has been said about a 42:03lot of these AI search features, right, whether 42:05it is perplexity or what Google is doing. 42:08is that increasingly they're kind of 42:09moving to a world where you don't have 42:11to go to the underlying website, right? 42:13It kind of curates the result for you. 42:15And so I guess Chris, to your original question, 42:17it's like, it's, it's kind of very weird to see. 42:19It's almost like the whole feature 42:20has been turned inside out, which 42:22is you are going to a browser bar. 42:24To not browse the web, but instead 42:27to get the results of a chatbot. 42:29That's, that's really 42:30strange, from my perspective. 42:31And that chatbot is gonna launch its own 42:34browser instance, and then Google somewhere, 42:37go to somewhere else, look up that website, 42:39and then come back with the answer. 42:40It, it's gonna be weird. 42:41Definitely. 42:42And I think there's almost kind of like, uh, 42:44the, the interesting debate here is I think the 42:46business rationale for Perplexity makes sense. 42:49It's almost a question though of 42:50like, how valuable is that browser 42:52bar going to be in the future? 42:53It of course has been like, I mean, it's 42:55been the source of a lot of litigation 42:56on like, say, Apple working with Google 42:58to have it as the default search engine. 43:00But you can almost imagine a future 43:02where, you know, maybe the app is 43:04actually the more powerful thing. 43:05Like when I go to want to know something, I 43:07actually don't go to the browser bar anymore. 43:09I just go to perplexity. 43:11Or in Kaoutar's world, it's like, 43:13I just speak into my phone, and the 43:14phone just does what I want it to do. 43:16Um, you know, I guess, uh, Vyoma, maybe I'll 43:19ask you, is like, is there a world 43:22where almost like, uh, Perplexity is trying 43:25to seize this real estate on the phone, which 43:28actually might not be so valuable in the future? 43:29Like, maybe browser bars are just 43:31like, going to be a thing of the past? 43:32Just FYI, uh, it, it has taken 43:35over my browser bar for sure. 43:36I've been using Perplexity for months now. 43:38And so I have many of my friends, you 43:40want to research anything, like in my past 43:42days, I would go back and like be like, 43:44Hey, I want to order these new headphones. 43:47space, Reddit. 43:48Now I no longer have to do that. 43:50Like, I, my, it's on my home screen. 43:52It's right there on my mostly used 43:54apps, because that's all I use now. 43:56I don't research anything no more 43:58about, so it has totally taken all that. 44:00And AI layered integration at the OS 44:03level is much better than anything, 44:05any standalone app that exists. 44:07So I think Perplexity has hit it outside 44:09the park there, and the only one thing that 44:12I sometimes struggle with in Perplexity 44:15is when I actually want to buy something. 44:17So let's say I want to buy a mattress or 44:18like a particular lampshade, then I will 44:21go in and then I'm researching about it. 44:23It's not getting that kind of consumer 44:26knowledge about me, like Google has, 44:28because of course Google's integrated 44:30at OS levels for that context for years. 44:33So this is going to be a great pivot for 44:35them to make their product or their large 44:38language model or their app more context 44:41aware, which is the need of the hour now. 44:43So it's going to feed all 44:45these usage patterns back. 44:47And I think I'm never going to 44:49lose perplexity from my phone. 44:51I love it. 44:51Genuinely. 44:53I no longer have Google search 44:55bar on my phone anymore. 44:57That's a, that's a big deal. 44:58Yeah. 44:59And, and, and I live in the Bay Area and 45:01there are many, many people who use that. 45:03So believe me, like you'd see them pop 45:05it out on their phones all the time. 45:06I have friends who use that. 45:08So I feel that is, um, one of the things 45:10that I'll see, but the one thing that they 45:12didn't speak about in that entire blog 45:14post was the hardware specifications of it. 45:18So what is that inference, um, 45:21layer that they're going to use? 45:22Is it going to be sound? 45:23Robust local interference with like the 45:28typical cloud or are they going to be 45:30using on prem such as the NPUs or TPUs, 45:32etc. So that is going to be the deciding 45:36factor, whether it is here to stay or not. 45:39That's a very good point, Vyoma, because this 45:42definitely depends on how mature or how powerful 45:45the edge AI models, especially the on device AI. 45:49So, as we're getting, uh, basically these 45:52LLMs to become smaller and more efficient, 45:55more AI processing can be done on the device. 45:58And this provides, of course, faster responses 46:01and better privacy. 46:02And also what you did is context 46:04that allows customization. 46:06So I, I see this evolution. 46:08I go in hand in hand as the edge AI becomes 46:11really more mature and more powerful. 46:13We can do more with these AI phones. 46:16Exactly. 46:17Yeah, the price angle I think is, 46:18I hadn't really thought about that. 46:19That's, that's very interesting is, you 46:22know, almost how, how cheaply can you pack 46:24these features into the phone is going to 46:27be this really, really interesting question. 46:28Cause right now it's almost 46:29like a luxury feature. 46:30If you want to run kind of a more sophisticated 46:32model, then we got to have all the power and all 46:34the build out and all the hardware at the edge. 46:37It just makes for a much more expensive phone. 46:38And they're promising for like 46:40less than a thousand dollars. 46:41And again, it's already very funny 46:42to be like, it's a phone, but. 46:44It's going to be cheap. 46:44It's going to be less than 1, 000. 46:46But even still, like, I think to kind of pull 46:48that off is like, pretty interesting in terms 46:50of how far you can democratize this tech. 46:52Yeah, it is going to start this whole 46:54wave of having these specialized devices. 46:57I hope we are not going back to the era of 46:59the Amazon, Amazon Firephones, how it kind 47:02of detached itself and wasn't that great. 47:04But I hope that this kind of breaks 47:07that curse and we are able to see 47:09something greater and better on this. 47:12I agree Vyoma, I think I just, I want to see 47:15that new experience as you say there, right? 47:17And you know, yeah, a native integrated 47:20experience and, you know, have all your 47:23contacts but it be private and, you know, and 47:26I loved your point, Kaoutar, about voice, etc. 47:28I think there's so many different modalities 47:30that can kind of come into this and, you 47:32know, and again, back to the camera as well. 47:35So I just, I just hope that we get 47:37a different and new experience. 47:39 47:40But I, as I said earlier, right, is if 47:42you want to, if you want to control that 47:44search experience, you need a device there. 47:47So I, I do think it's a brilliant move. 47:50Yeah, this is all happening while 47:52Apple is delaying its AI features. 47:55I don't know your take on this. 47:57Uh, is that because they want to make sure 47:59that they have a very well curated, secure, 48:03because Apple in general has been conservative 48:05about, you know, the security, Or, you know, 48:09this is opening the space for more, for 48:11example, for Perplexity and so on to take over 48:15some of the market that, um, Apple phones have. 48:17Yeah, I think that the Apple part, when they 48:20put out Apple intelligence and they hear, like, 48:22they got a lot of backlash about that as well. 48:25So maybe they are field testing it way, 48:27way more before coming into production. 48:28I don't know if you know about this, but that 48:32Apple, uh, came up with the Apple Kids Watch. 48:35Because again, they're as it is like the 48:37point that you made, Kaoutar, they're known as 48:39the company which respects privacy and has 48:42it integrated in all aspects and they've come 48:45them coming up with the kids watch kind of 48:47showcases their, um, commitment towards it. 48:50So I feel they are looking into several 48:53avenues before coming out with something. 48:55Public 48:55yeah, and I still think I think weirdly the 48:58kind of hardware mentality actually might 48:59be working against them a little bit in 49:01implementing some of these features because 49:03sort of unlike you know building a phone which 49:05you can really kind of like I feel like part 49:08of the problem with these models is they still 49:10You know, unreliable and kind of probabilistic. 49:13And, you know, I think like in some 49:14ways the discipline of like launching 49:15features is a little bit more risk 49:17loving than that Apple might be used to. 49:19And I think it's actually 49:20holding them back in the market. 49:21Yeah, but I ain't giving up 49:22my iPhone for anything, Tim. 49:23So I'm okay with that. 49:26Yeah, that's right. 49:27I mean, I think the counter argument as well. 49:29Uh, they can just keep trying because 49:30everybody's on their phone and 49:32they're not going to throw it away. 49:33So they can just keep going until they get it. 49:35Um, but something to keep an eye on. 49:37Um, we'll definitely be keeping 49:38an eye on this Perplexity project 49:40and, um, a lot more to come there. 49:42Um, so, uh, that's all the 49:44time we have for today. 49:45Uh, Vyoma, Kaoutar, Chris, uh, 49:47thanks for joining us as always. 49:48Um, and, uh, thanks for 49:49joining us, all you listeners. 49:51If you enjoyed what you heard, you 49:52can get us on Apple Podcasts, Spotify, 49:54and podcast platforms everywhere. 49:55And we will see you next 49:56week on Mixture of Experts.