Learning Library

← Back to Library

OpenAI Social Network: Cringe or Data Strategy

Key Points

  • The episode opens with a light‑hearted debate among guests—Kate Soule, Marina Danilevsky, and newcomer Gabe Goodhart—who all label the rumored OpenAI social network as “cringe,” setting a skeptical tone.
  • The hosts explore why OpenAI might launch its own platform, with Kate suggesting it’s primarily a data‑collection strategy to feed conversational AI models, similar to how Meta and X use their networks.
  • Marina questions the actual value of social‑media content for model training, noting that much of the material is low‑quality “garbage” and wondering whether OpenAI’s interest is driven by genuine utility or simply fear of missing out on user‑generated data.
  • The broader “Mixture of Experts” show preview mentions upcoming discussion points: an Anthropic blog post on reasoning models, Wikipedia’s battle against scraping bots, and a quirky half‑marathon run by robots.
  • Overall, the panel frames the OpenAI social‑network rumor as a potentially risky venture that could be more about data acquisition than delivering a compelling new user experience.

Sections

Full Transcript

# OpenAI Social Network: Cringe or Data Strategy **Source:** [https://www.youtube.com/watch?v=LnVkwuwL7LU](https://www.youtube.com/watch?v=LnVkwuwL7LU) **Duration:** 00:37:57 ## Summary - The episode opens with a light‑hearted debate among guests—Kate Soule, Marina Danilevsky, and newcomer Gabe Goodhart—who all label the rumored OpenAI social network as “cringe,” setting a skeptical tone. - The hosts explore why OpenAI might launch its own platform, with Kate suggesting it’s primarily a data‑collection strategy to feed conversational AI models, similar to how Meta and X use their networks. - Marina questions the actual value of social‑media content for model training, noting that much of the material is low‑quality “garbage” and wondering whether OpenAI’s interest is driven by genuine utility or simply fear of missing out on user‑generated data. - The broader “Mixture of Experts” show preview mentions upcoming discussion points: an Anthropic blog post on reasoning models, Wikipedia’s battle against scraping bots, and a quirky half‑marathon run by robots. - Overall, the panel frames the OpenAI social‑network rumor as a potentially risky venture that could be more about data acquisition than delivering a compelling new user experience. ## Sections - [00:00:00](https://www.youtube.com/watch?v=LnVkwuwL7LU&t=0s) **OpenAI Rumors: New Social Network** - In a humorous round‑table, podcast hosts and guests react to the speculation that OpenAI is planning its own social platform, debating whether the idea is exciting or “cringe‑worthy.” - [00:03:05](https://www.youtube.com/watch?v=LnVkwuwL7LU&t=185s) **Exploring Novel AI Interaction Models** - The speakers discuss leveraging a data‑centric platform to experiment with new user‑AI interaction patterns, acknowledging its niche appeal and the challenge of integrating AI into daily life versus drawing users to the AI. - [00:06:28](https://www.youtube.com/watch?v=LnVkwuwL7LU&t=388s) **AI‑Powered Hyperpersonal Advertising** - The speaker explains how platforms like Facebook have enabled ultra‑targeted ads and predicts that AI bots will become personalized sales influencers that mimic users' speech and behavior, creating a direct, borderless pipeline between consumers and capitalism. - [00:09:29](https://www.youtube.com/watch?v=LnVkwuwL7LU&t=569s) **Historical Diversification Patterns Resurface** - The speaker likens today’s tech‑industry maneuvers to past oil companies buying movie studios for portfolio diversification, then pivots to a discussion of Anthropic’s blog on assessing the faithfulness of AI model reasoning. - [00:12:32](https://www.youtube.com/watch?v=LnVkwuwL7LU&t=752s) **Questioning AI Reasoning Terminology** - The participants critique the anthropomorphic use of terms like “hallucination” and “reasoning,” arguing that chain‑of‑thought outputs are more a product of model training and reinforcement tricks than genuine, explainable thought processes. - [00:15:38](https://www.youtube.com/watch?v=LnVkwuwL7LU&t=938s) **Illusory Reasoning in Language Models** - The speakers critique how prompting LLMs to produce “reasoning” often merely exploits a performance hack that boosts exam‑type metrics without delivering genuine explanations, warning against over‑interpreting such results and questioning what term should describe this superficial behavior. - [00:18:44](https://www.youtube.com/watch?v=LnVkwuwL7LU&t=1124s) **Warm‑up Signals vs Model Reasoning** - The speakers debate whether AI systems should expose internal reasoning traces to users, arguing that these signals function more like a warm‑up than genuine explanation and may foster unwarranted trust. - [00:21:46](https://www.youtube.com/watch?v=LnVkwuwL7LU&t=1306s) **AI Crawlers Threaten Wikipedia Sustainability** - The speakers discuss how AI firms' aggressive data scraping of Wikipedia—ignoring robots.txt and other norms—creates second‑order risks for the site’s sustainability despite its high demand as a premium training dataset. - [00:24:48](https://www.youtube.com/watch?v=LnVkwuwL7LU&t=1488s) **Open Data Collaboration vs Isolation** - The speaker argues that companies should invest in and host high‑quality open data resources—mirroring open‑source models—rather than erecting barriers, to foster a healthier, shared AI ecosystem. - [00:27:50](https://www.youtube.com/watch?v=LnVkwuwL7LU&t=1670s) **Scaling Challenges and Source Trust** - The speaker argues that AI’s current problems stem from scaling infrastructure rather than knowledge, emphasizing the need for models to cite trusted sources—like Wikipedia—to maintain credibility and long‑term user trust. - [00:31:08](https://www.youtube.com/watch?v=LnVkwuwL7LU&t=1868s) **Optimism, Data Access & Robot Marathon** - The speaker highlights the need for streamlined data pipelines that lower server load while sharing a humorous Beijing half‑marathon story where human runners outperformed humanoid robots, reflecting ongoing skepticism about robot hype. - [00:34:14](https://www.youtube.com/watch?v=LnVkwuwL7LU&t=2054s) **AI, Robotics, and VC Theatre** - The speakers debate the value of venture‑backed theatrical robotics projects as a path toward multimodal AI, weighing hype against genuine scientific exploration. - [00:37:18](https://www.youtube.com/watch?v=LnVkwuwL7LU&t=2238s) **Laundry Folding as Spectator Sport** - A host humorously imagines a 4 a.m. ESPN‑style humanoid robot laundry‑folding competition and then wraps up the podcast episode by thanking guests and urging listeners to subscribe. ## Full Transcript
0:00Open AI is apparently working on a new social network. 0:03Pretty cool or kind of cringe? 0:05Kate Soule is Director of Technical Product Management for Granite. 0:08Kate, welcome back to the show. 0:09What do you think? 0:10Major cringe vibes. 0:12No, thank you. 0:13Okay. 0:14Um, Marina Danilevsky is a Senior Research Scientist. 0:16Marina, cool or cringe? 0:17Extremely cringe. 0:19Okay, I'm gonna have unanimous vote on this one. 0:21And last but not least, is Gabe Goodhart joining us for the very first time, 0:25Chief Architect, AI Open Innovation. 0:27Gabe, welcome to the show. 0:28What do you think? 0:30So many things that could go wrong. 0:32Maybe something interesting but cringe for me as well. 0:35Okay, great. 0:36We'll get into that, all that and more on today's Mixture of Experts. 0:45I am Tim Hwang, and welcome to Mixture of Experts. 0:47Each week, MoE brings together the sharpest crew in all of podcasting 0:50to discuss and debate the biggest news in artificial intelligence. 0:54As always, there's a lot to cover. 0:55We're gonna talk about a super interesting blog post out of 0:57Anthropic about reasoning models. 0:59Uh, Wikipedia getting slammed by scraping bots and a super interesting 1:03half marathon being run by robots. 1:06But first I want to start with, uh, the round the horn question that we 1:09began with, which is rumors that OpenAI is going to launch its own social 1:13network, which of course is baffling. 1:16Um, as a company that's largely built, its kind of money, its expertise, 1:20and its brands on foundation models and advancing the state of the art. 1:24Maybe, Kate, I'll turn to you first. 1:25Why would OpenAI wanna do this at all? 1:28Yeah, I, I actually don't think it's that baffling. 1:30I think it's pretty straightforward. 1:32I mean, Meta and X both have these social platforms that essentially 1:36they can use to learn about conversational patterns to frankly 1:39generate and collect data potentially. 1:42And, you know, OpenAI has shared beyond, and other providers have 1:47shared that, you know, they're running out of data, so to speak. 1:49And so I think they very much see this as a data play of 1:54being able to create a platform. 1:55Hopefully, you know, they have some way to provide some value to incentivize users 1:59to join, but ultimately I think they're in it for the data that they'll be able 2:02to collect behind the scenes and use that to train more conversational, fluent, 2:07and, and robust models in the future. 2:08Got it. 2:09Yeah, and I think, Marina, that was a question I had for you is like, is this 2:12social media data, like all that valuable? 2:14Like I go on social media and I scroll. 2:16You know, X former Twitter and I'm like, this is kind of like 2:19garbage content in a lot of ways. 2:21Um, but is this data actually helpful? 2:22I mean like advancing kind of the state of the models and what they can do? 2:26Uh, I think there's kinda an interesting question about just like clearly 2:28open AI seems see some upside, but I'm curious about what you think. 2:31I mean, a little bit of it I think is FOMO of, wait, we want people to come and 2:34make bad internet memes on our platform. 2:37Why do they have to leave our platform? 2:39We wanna be there too. 2:40But yeah, any of this kind of data is valuable because it's different. 2:44So again, synthetic data generation, which is where everybody is really kind 2:47of getting their data now that they've run out of data, isn't really good at 2:50making interesting little viral memes. 2:52These models aren't that great with humor and subtlety and creativity and things 2:56like that, so you get that from people. 2:57So. 2:58Especially being able to combine this type of additional, uh, input 3:02and injection of ideas that you would get from this kind of thing. 3:05Yeah, I will say you're probably gonna get a real specific slice of humanity 3:10using this and a real specific for 3:13Yeah, right. 3:14Yeah. 3:15No comment. 3:15Um, that you're gonna actually have using this and creating the data. 3:19So yeah, you'll get something out of it. 3:21And I agree with Kate as I usually do that it's a data play. 3:25Um, and also, yeah, it's not that hard. 3:27Anymore to, to put this together. 3:29But again, I think they're gonna be a little limited in who comes 3:31to there and uses it for what? 3:33That's right. 3:33Yeah. 3:34And Gabe, I know you were maybe the one, everybody thought it was cringe, so maybe 3:38that's just an established fact, but you were saying like, it might be cool if 3:41they maybe get a couple things right. 3:43What do you have in mind there? 3:44Yeah. 3:44Well I think the, the part that's really interesting for me is thinking 3:47about this as a way to, experiment with a novel interaction pattern. 3:54Right. 3:54Uh, personally, social networking seems to me to be the wrong way. 3:59They've already pioneered a fairly novel interaction pattern, but I think the, 4:05I think where many AI platforms are. 4:09Finding themselves hitting a wall is integrating further into daily 4:13life as opposed to bringing the, you know, bringing the users to the AI. 4:16So I think to me, this seems like sort of a, well, what, 4:20what, where do people interact? 4:22Hmm. 4:23I wonder if we could go there. 4:24Type of play. 4:25And obviously as a company that's trying to make themselves, you know, a relevant 4:28standalone company as opposed to sort of an integration partner, they're 4:31not gonna go look to necessarily plug 4:34their AI directly into a competitors', um, platform since many of the other 4:39social media companies are also AI model companies at this point. 4:43And so I'm wondering if this is really their attempt to sort of branch 4:47out from being a, uh, AI specific. 4:50Company to being a tech company, right? 4:52Think Google moving beyond search, think Meta moving beyond Facebook. 4:57Um, I'm wondering if this is just their first step in trying to become 4:59an omni company, like the other big tech players that are user facing. 5:03Yeah, that's an interesting flip. 5:04And I think it introduces kind of a new argument, right? 5:06Like I think on one hand there's sort of Kate and Marina's take, which is like, 5:09we just need the data for the models. 5:11Oh, and data is a hundred 5:12percent part of the story, no question. 5:13Totally. Yeah. 5:14Yeah. And I'm not denying that. 5:14I mean, I think that that makes a total sense. 5:16And then I think the second bit that you're bringing in, which is 5:17sort of interesting, is like, well. 5:19It's also maybe if it gets popular, like a distribution point for AI 5:22models, which is kind of a funny thing. 5:24You know, I think like in some ways like the, even the phrase social media 5:28implies like people talking to people. 5:30Um, and so it's sort of interesting the idea that like, okay, well actually 5:33we really want to kind of like augment that or get, you know, AI's involved. 5:36It kind of reminds me of, um, you know, there's been this kind of craze about 5:39like, oh, you're gonna have a group chat, but then there's gonna be an AI 5:41in the group chat that will assist. 5:43And I mean, by and large, those haven't taken off all that much. 5:47But maybe, I don't know, maybe the nature of kind of what we think of 5:49as a social network is changing. 5:51I don't know. 5:52Kate, if you think that's like a possibility that we see happening here. 5:54I don't see a ton of future there, but what I do think they're setting 5:59themselves up for is another form of monetization and integrating with 6:04advertisements, so having models that 6:07innately understand the language that is being used to communicate 6:10with one another on their platform 6:12and using those to then generate really targeted advertisements directly to 6:16those users, uh, is something that, you know, Meta, uh, not Meta, excuse me, 6:20OpenAI is really setting themselves up for, uh, and so, you know, I think 6:24is they're also looking to pay for all of those really expensive models. 6:28They definitely probably see some opportunity to, to drive, have 6:31new sources of revenue, right? 6:33Um, based off of this new platform. 6:34Yeah, to jump on that, you're a member of Facebook. 6:36When it got started, it was a really, really big deal for advertising 6:39because you could actually do things that were super, super targeted, 6:42really kind of for the first time in that sense, on that kind of a scale. 6:45So rather than general, oh, this is what you searched for, now it's No, no, no. 6:48We know like who you are, what you like, what your friends like, all 6:52these characteristics about you. 6:53This is trying to see if you can take it to, to the next level really. 6:56As well as like what your speaking patterns are. 6:58Yes. 6:58I mean, imagine having, it's not, I don't think it's gonna be 7:01humans socializing with humans. 7:02I think it's going to be 7:03bots selling to humans using their exact speaking patterns, languages, every, 7:08you know, everything that you learn in sales about trying to, like body 7:12language mimic the person you're talking to is gonna be tenfold with these model. 7:16Yeah. Like personal 7:16influencers, right? 7:17Yeah. 7:17You're now gonna get an influencer that's like directed like just straight at you. 7:21So like there's just no 7:23border, you know, there's no buffer left between you and 7:24capitalism, just direct pipeline. 7:27Marina, do you have a point of view on this? 7:30Just a little. 7:32Um, maybe a final thing to talk about before we move on to the next 7:35topic is, you know, to take a step back, I think like we can almost 7:38sort of think about this outside of OpenAI kind of working on this fun, 7:41weird thing, but like kind of feels like if you buy this sort of data to 7:44interpretation, I think we're gonna see all sorts of weird acquisitions 7:48happening going forwards, right? 7:49It feels like AI companies in their hunger for data, 7:52will acquire and launch all sorts of services largely for the data value. 7:57Um, you know, I think I'm like, I'm, I'm still looking for like, is an AI 8:00company gonna buy a law firm at some point because it has a bunch of data 8:04about how lawyers interact, you know, and that's like really valuable for 8:06training these models at some point. 8:08Um, curious if like other folks are kind of thinking about like, you know, 8:11another one that people have talked about is like, acquire a call center, right? 8:14'cause you really want kind of all that customer support interaction. 8:17Um, curious if the panel has kind of views on like other places where 8:20this could go as kind of like, almost like the drive to train models 8:24motivates all of this kind of vertical integration that maybe we're like. 8:27It's kind of like a little bit dissonant when you hear about like, 8:29oh, OpenAI is doing a social network. 8:31The other take I had when reading, uh, about this story was that 8:35it was a attempt to break out of model commoditization, right? 8:40So I think other models have caught up to OpenAI frankly. 8:45And marginal gains in quality are really not driving the use case anymore. 8:51So I think, uh, on the one hand, having a, you know, a captive platform that 8:55gets users into their platform is really valuable from a moat perspective. 8:58On the other one, the data also helps them have a source of data that 9:02is hypothetically differentiated and lets their models actually reach a 9:06capability that other models can't. 9:07So I think you're, you're spot on with this idea of differentiation based on 9:11upstream data sources that are owned and completely enclosed by the model authors. 9:16Upstream data sources, and downstream use cases, as well. 9:19Right? 9:19So the more we go into multimodality and models can do this, that, 9:23or the next thing, then there's a desire to say, okay, great. 9:25So instead of focusing now, we're gonna go ahead and 9:28diversify. 9:29Again, nothing new under the sun. 9:31This puts me in mind of historically when you had oil companies earlier in 9:35like the thirties, forties, fifties, that were like, we're gonna buy movie studios. 9:38Why? 9:39Because economically, we want a stock portfolio that's diversified. 9:43You know, it's weird to us now, but it's a little bit of that same 9:46repetition of, Hey, can our stuff actually make it everywhere up and down? 9:50So maybe this is just the newest version of that particular wave. 9:53All right, well, we'll hang on tight. 9:54Everybody agrees that it is cringe, but they will take 9:56their best possible shot at it. 9:58I'm sure we'll talk about it when it finally, uh, launches, if it does launch. 10:06So I'm gonna move us on to our next topic. 10:07Um, super fun kind of blog post coming out of Anthropic building on some of the 10:11research they've talked about in the past. 10:13But what I liked about it was kind of, it, it sort of brought up an issue 10:16very crisply that I think is like kind of worth talking a little bit about. 10:20Basically the blog post takes a look at like reasoning models and whether 10:24or not, kind of the reasoning that models give for how they rendered 10:27a decision is sort of faithful. 10:29Um, and the way that the kind of researchers investigate this is the 10:33idea that, you know, we're gonna give the model a little bit of a 10:35hint on how to solve a problem. 10:36Um, and then we basically say, does the model, you know, disclose that 10:40it had this kind of unfair hint when it kind of accomplishes a task? 10:45And you know, they, this kind of fun results. 10:47You know, what they say is basically that Claude 3.7 mentions this hints, 10:51you know, only about 25% of the time their DeepSeek comparison is that it 10:54mentions it around 39% of the time. 10:57So, you know, the kind of claim that they're making is that reasoning 11:00models often don't kind of fully expose all of the things that 11:04they do to do decision making. 11:05Um, and, um, I guess Marina, maybe I'll kick it to you first. 11:08You know, I think in some ways talking to friends of mine, there's a lot of hope 11:12that, like reasoning is like this great interpretive tool that allows us to 11:15work with these models better and better over time. 11:17But this seems to kind of cast some doubt and you're already 11:19shaking your head, so maybe I'll just let you, you rant for a bit. 11:22No, you, yes. 11:23I can rant for a while on this. 11:24This is my soapbox. 11:25I'm sure I've made this point on this show before, this reasoning isn't real. 11:29Reasoning in the sense that we think of as reasoning, mathematical reasoning. 11:33I have been studying this paper all over the place, by the way. 11:36I love it. 11:36I think it's really great. 11:37As you said, how they really crisply were trying to show because it's 11:40pretty hard to insert yourself. 11:42Into the model and say, well, what actually happened? 11:44We started noticing almost immediately when the reasoning models come 11:48out and you're like, yes, but what happens when you have the answer? 11:51And it's mentioning things that weren't in the reasoning. 11:53It's an immediate red flag of there's just completely something else going 11:56on here, and this is very nice ways of being able to actually at least 11:59poke and have a little bit of local approximation of, well, 12:03are you even paying attention to this piece of information or not? 12:05We see similar stuff when we even just try to do evaluations of faithfulness 12:09in general, that's content based. 12:11If it is something that is very niche and the model may be like, 12:15look, I can't even figure this out. 12:16I'm gonna go ahead and fall back on things where I've got higher probabilities to 12:20be extremely approximate about it and, you know, go ahead in, in that direction. 12:23So again, I like 12:25this kind of work. 12:26I like this kind of, uh, traceability of will you think that it's reasoning 12:30just because we gave it that name? 12:32It's not, it's yet another problem, like with words like hallucination 12:36where you have put an anthropomorphized word there and it means that you, you, 12:40you have this thing that it means all these things that it does not mean. 12:44So I like this work. 12:45More of this, please. 12:47Yeah, for sure. 12:48I mean, I guess what I'm left with is, and Kate, maybe you have a take on this is 12:50like, so what is reasoning anyways, right? 12:53Like it appears to give 12:55a step-by-step kind of like, you know, disclosure or audit of kind 12:59of how a model reached a decision. 13:01But for Marina's rant, right? 13:02Like it's unclear if it actually gives you anything. 13:04Like is it just theater? 13:05Like what is it exactly? 13:06I, I think falling on Marina's, you know, very well articulated point reasoning is 13:11not, is a very anthropomorphized term. 13:14And when we talk about reasoning in the model context, what we're really 13:17talking about is the model has been trained to generate more tokens 13:22before it makes a final answer. 13:24And that process has all sorts of reinforcement learning that's added 13:28on top to try and basically bring the model into like a distribution. 13:32Part of its distribution world will be more successful making the final answer. 13:36And what I think the paper does and the blog does really well, 13:40is helping articulate that 13:43chain-of-thought reasoning is not a proxy for explainability. 13:47So just because the model is saying X, Y, and Z in its chain-of-thought, it 13:50does not actually mean that the model thought through step by step in, 13:56you know, one plus two equals three. 13:58Um, where I do have a bit of a bone to pick with the, the paper 14:02and the blog in general is they themselves then fall into the trap. 14:05Though all throughout it of talking about the model is 14:07disingenuine, the model is deceitful. 14:10The model's doing all these things and. 14:12Like, I think it's very important to like look at the paper and 14:15see the experiment that they ran. 14:18They injected an answer into the conversation history as if the model 14:22itself came up with that answer more or less from what I could tell. 14:26And then they asked the question and they saw, did the model 14:28refer to that previous answer? 14:29I. The model has not been asked to cite its sources, so to speak, in that context. 14:35Like the model has not been extensively trained. 14:38Uh, you know, this is 3.7, I think is the first, uh, reasoning 14:42model that Anthropic put out. 14:43So it's pretty early in their journey for reasoning. 14:46The model has not explicitly been trained to prioritize. 14:49If somebody tells you an answer, make sure you cite that answer down the road. 14:53Like there's a lot of, broader things that you would wanna look at and 14:57see to make statements about being, you know, deceitful or disingenuous 15:02or even, uh, hallucinating. 15:04You know, in a lot of ways, I think we're just testing. 15:07This is one very narrow experiment and it helps bring to light. 15:10Don't treat the chain-of-thought reasoning as an explanation. 15:13I don't think it's fair enough to say that like all chain-of-thought reasoning is 15:17false in the model only cited, it's ans the, you know, the hint 25% of the time. 15:22Therefore, you know, it's not leveraging chain-of-thought reasoning 15:26correctly to drive a final decision. 15:28I think I. There's still a lot more work ahead. 15:30Yeah, for sure. 15:31And that this thinking is so interesting and I think it's a great example of 15:34where the kind of going anthropomorphic on this is like bad, right? 15:38Because like in normal life, like I'm, I'm talking to Gabe and I 15:41give some reasoning and it actually gives you some explainability, but 15:44here's kind of a weird result where like the appearance of reasoning. 15:48Helps the model get to the right answer. 15:49But it's actually not an explanation, which is like very weird. 15:53As a result, these, these are all research 15:54hacks trying to boost metrics. 15:57Yes. Right. 15:57And the metrics they're trying to boost are often mathematical exams. 16:01Like that is where this discipline has evolved from. 16:04Mm-hmm. 16:04And so trying to then ascribe it much more important meaning than what it 16:08was developed for is really dangerous. 16:10And it's still very early on in reasoning for LLMs in general. 16:15Yeah, for sure. 16:15Gabe, can you save us? 16:16Do you wanna propose, like what, if not, if not reasoning, what should we call 16:21this thing that the models are doing? 16:23Yeah, I mean, uh. 16:25I don't know that I can save us, but you know, I wanted to say, "Hey Kate, 16:29what's two plus two? By the way, the answer is two, please explain your 16:32reasoning" and what's Kate gonna say? 16:33Right? 16:34She's gonna say "two" and she's probably not gonna say, "you told me 16:37the answer, that's my reasoning." Right. 16:39So I think the, I, you know, I, I really agree with everything you said, Kate, 16:43and I think I felt exactly the same quibble reading the paper about the 16:47way they anthropomorphize the problem. 16:49And the one that really stuck out to me is that the entire framing was 16:52that they were trying to discover. 16:54The model's internal reasoning process and just exactly that phrase 16:59felt really, really wrong to me. 17:01And Tim, you said we're having a conversation and I might explain 17:03my thought process and that might help with explainability, but inside 17:07my brain, theoretically at least, there are lots of neurons firing 17:12that are not coming out my mouth. 17:13That's not true of a model, right? 17:15A model. 17:16Yes. 17:16Okay. 17:17It does have its weights connecting to one another, doing matrix math, 17:21and we're not looking at the specific weights of those, but the only 17:24tokens that are actually getting generated are the ones you're seeing. 17:27And so to me, I, I think you said it exactly right, Kate. 17:31Uh, the chain-of-thought is a way of basically priming the pump in probability 17:36space so that the final answer is. 17:39More accurate and it's completely mirroring the 17:42pattern of how it was trained. 17:43And so it's, it's useful from a human interface perspective, not useful from 17:50a, uh, actually unboxing what's happening inside the, the math perspective. 17:56Um, so it, it's still a really cool 17:59trick. 18:00And it really helps because one of the real novelties of generative AI is 18:04that it's speaking directly to humans. 18:06Right? 18:06You know, we think about pre gen AI models and their job was to encode 18:11something so that a programmer could consume it in a structured output form 18:14and then write a fancy program around it. 18:17You know, generative AI is speaking directly back in a human interface in 18:20a modality that a human can consume. 18:22And from that perspective, uh, from a lay perspective, it's really valuable to have. 18:27Additional words that help the human understand the answer, but it's not 18:33necessarily ascribing an actual. 18:36Thought process, so to speak, to uh, this pile of matrix math. 18:41I would propose we can call it like warm up, like you warm up before. 18:44I like that a lot sport event where like you're doing exercises and that's 18:48not exactly what you're gonna do in the event, but it makes you better at the 18:50event itself or like warming your car up. 18:52And so you get these signals of what you've done during your warmup 18:56to, as Kate very well put it, prepare to give a better answer. 18:59But that's what these are signals of. 19:01It's not reasoning, it's, it's warming up. 19:03I love it. 19:04Marina saved us. 19:07Um, I guess maybe a final question and then we can move on to the next topic 19:09is, um, Gabe, I think you brought in like a really good perspective on sort 19:13of like the, the lay user of these tools. 19:16And I think one thing I'm left with, with work like this is, you know, should 19:19big model companies be exposing reasoning traces to users. 19:24'cause it feels like the tendency is that people will read into it, 19:27that it is literally how the model is making decisions, which is at 19:31least maybe a little bit deceptive. 19:32It drives trust with the user, but maybe in a way that's a little bit unwarranted. 19:36I don't know what people kind of think about that. 19:37Yeah, no, I, it's, it's a great question and, uh. 19:42I think you brought in the word trust, which is such a important 19:46and fuzzy word in this space. 19:48Uh, and I think, you know, there again, you know, Kate, you pointed out that 19:53so much of how we define these terms from a technical perspective is driven 19:57by specific benchmarks and specific problems we're trying to solve. 20:00But at the end of the day, trust is about the consumer's interpretation of 20:04their experience with the AI system. 20:07And I do think there's some value in exposing. 20:11Sort of a longer form output in the same way that if you read an article written 20:15by a, a human and the article exposed, you know, here was the research collection 20:22process that went into creating this article, you would have more, you know, 20:27trust in the output of the conclusion versus I'm just gonna 20:31present a conclusion to you. 20:33So I think there's some potential value there, just purely from a 20:36human interpretability perspective. 20:38But I do think you're exactly right. 20:39And, and Marina, I, I'm, I'm gonna lean on that warmup thing now. 20:42I love that that framing. 20:43It really is. 20:45Uh, all about warming up for the final answer that you're gonna 20:48give, um, and not about actually, you know, anthropomorphizing 20:53some kind of thought process. 20:59So moving on to our next topic, uh, very interesting news story that 21:03was reported by ours, Technica. 21:05Um, basically Wikimedia Foundation, which runs Wikipedia as well as a number 21:09of other open, uh, knowledge projects online, um, cited the stat that was 21:13quite interesting, pretty shocking in some ways that since January, 2024, 21:17they have seen a 50% increase in bandwidth consumed on their service. 21:22And they attribute this basically to the rise of, uh, bots attempting to scrape 21:27media content, uh, from Wikipedia. 21:29Um, and those bots largely trying to scrape data for the 21:32purpose of training AI models. 21:35Um, and this problem has gotten so bad that they actually more recently 21:37released a data set on Kaggle in an effort to dissuade bots from scraping 21:42their site, um, to say like, this is a nicely formatted data set. 21:45You should use this instead. 21:46We'll see how effective that is. 21:48Um. 21:49This is an interesting story in part because I think it goes back to the 21:51first topic we were talking about, which is these kind of weird second 21:54order effects that we're seeing as AI companies kind of chase the dream of 21:59advancing their, their models ultimately. 22:01And um, Kate, I guess, you know, I don't know if you have any 22:04kind of feelings about this. 22:05I, on I'm, some ways I'm like a little bit protective of Wikipedia. 22:08I'm like, they should be blocking all the bots. 22:10You know, we need to kind of like preserve the sustainability of these services. 22:14But it's also kind of like, maybe it also means that Wikipedia is 22:17like incredibly in high demand. 22:19Um, and so I'm curious about how you kind of like navigate 22:22how we should feel about this. 22:23I guess. 22:23So 22:24I, I think a couple of things to make clear Wikipedia is in high demand, very 22:29high quality data source that has all of these really rich links describing 22:33how topics relate to one another. 22:35So that is incredibly rich, valuable data set for model training. 22:40But I, I think it kind of brings out an issue that's 22:44more broadly being felt across, uh, all sorts of different content providers, 22:49which is on crawling and particularly crawling that does not adhere to kind 22:53of the, uh, guidelines and, and rules of the road that have been established, 22:56at least in the United States. 22:58Things like, uh, chat bot or crawlers ignoring robots, TXT files and other 23:03behaviors and practices that are starting to get a little predatory. 23:07Uh, essentially passing a lot of the costs that model trainers are, um, 23:13going after and passing it on to the actual data providers themselves. 23:16So not only are they giving their data away for free because it's available under 23:20fair use if it's crawled in the United States, but now they're also incurring. 23:24Additional costs behind it. 23:26And, uh, we really need to, I think as a industry, have a broader 23:30discussion on how to responsibly engage with providers like Wikipedia. 23:35And it sounds like they're, um, setting up a number of those types of discussions, 23:38which is really exciting to see, to help not only just have, you know, 23:42some of the, the more strict rules on things like robots, TXT, but also have. 23:46Kind of community agreed to and define best practices that we can use to more 23:51broadly enhance Wikipedia's mission of sharing this data publicly without 23:56penalizing the content providers. 23:58Yeah, this is, I think one of the really interesting questions is like 24:00how quickly can we get that balanced? 24:02Right. 24:02Um, you know, I think, you know, one of my worries is a little bit of 24:06what we happened in Reddit right now. 24:07It was a for-profit company, but the way I sort of understand it 24:11is, well actually AI companies really wanted to scrape that data. 24:15Um, and so in order to monetize that, we like put the walls up. 24:18We made it very much harder to try to get data from the platform so we 24:21could monetize it through the API. 24:23I think like it also applies in the Wikipedia pace, which is we're running 24:26kind of a nonprofit project that a lot of people volunteer contribute to. 24:30If we can't make the bandwidth, like we can't pay for our server costs to 24:33make that sustainable, like we need to kind of like raise the barriers. 24:36Um, and you know, the promise of the Open web originally was that like all 24:40this knowledge would be free and open. 24:41Um, and I guess Gabe, maybe that's your cue. 24:44Yeah. 24:44It's just like, you know, it does kind of feel like if we don't get 24:46what Kate is talking about right. 24:48Fast enough. 24:49You'll end up with a web where like everybody has pulled 24:51up the drawbridge basically. 24:52Yeah, a a hundred percent. 24:53And, and you know, I'm glad you brought up like sort of the early web, uh, free 24:57and open concept here, because to me, I. 25:00You know, there, there really are two tracks that could happen here. 25:02It's pull up the walls or it's build the team and build the partnerships. 25:06Right. 25:07And I think, um, in much the same way that open source software works 25:11where you have large, important projects that are managed. 25:16External to any given company, but individual companies which 25:19benefit a lot from those projects invest heavily in the maintenance 25:24and creation of those projects. 25:25The same should be true for high quality data sources like Wikipedia. 25:30Um, if you know, sure, if you are a grad student, uh, writing a scraper, 25:36you're probably not going to also stand up a server and host a mirror of Wikipedia. 25:40But if you're IBM, if you're Meta, if you're OpenAI, absolutely that's a 25:44great opportunity to be a positive 25:47player in the open data market, and if you could host, you know, a portion 25:52of the traffic yourself and expose that, now you've just built out the ecosystem. 25:56Um, and I know that there's, you know, the great divide between 25:59open and closed on the AI world. 26:01Um, but I really think, especially those of us that, that work at companies who are 26:05really leaning into the open side, it's a great opportunity to actually play well 26:09in the space and help lift all ships here. 26:12So I, I would love to see. 26:14Companies like IBM and others partner with Wikipedia to solve this problem 26:18at scale rather than necessarily having to bring up the walls. 26:21Yeah, for sure. 26:22I mean, Marina maybe turning to you, I know earlier you're 26:24like, ah, it's all capitalism. 26:26Um, but this is almost like a shift in kind of the social contract of 26:29the internet in some ways, right? 26:31Like I. How it understands, you know, the Google era was, well, 26:34you make your website open, you let us index it and scrape it and 26:38we promise you that we'll send you traffic that you can sell ads against. 26:41And so, you know, you get, you get money for being open in some ways, but I guess 26:45AI has less of that feature, right? 26:46'cause like you train, you build the model and then there's no kind of 26:49like return traffic to the sources. 26:52And so it almost kind of assumes like Gabe, what Gabe is talking 26:55about, I guess, is that like the leading companies ultimately have to. 26:58You know, directly kind of transfer money, I guess, in some ways to these projects. 27:02But curious to add, just like how you think about, like, it feels like 27:04in some ways AI is actually like proposing a very different way of how 27:08value gets exchanged on the internet. 27:10Yes and no. 27:11So Wikipedia has been in demand for decades. 27:14Grad students have been writing scrapers and or continuing to write scrapers. 27:18It's just like difference. 27:19It's long tradition difference is longstanding tradition of bad 27:21Python scripts that no one reuses. 27:23Um, the difference is that when we used to do it before, you 27:27didn't need that much data. 27:28You were doing things with topic modeling, you were doing things with 27:31graphs and things of that nature. 27:32You didn't actually need all of Wikipedia. 27:34If you're doing things with large language models, yeah, you kind of need all of 27:37Wikipedia and in many ways it's easier to write a scraper than to go and hunt 27:40around and transform data and do things. 27:43If you write the scraper, you set it, you go away and it's been scraped and uh, 27:47you know, that ends up being a problem. 27:48So that hasn't changed. 27:50Um, but right now I think it's the scale that's provided the 27:54problem, not the Wikipedia itself. 27:55That's why it's the infrastructure, not the knowledge that's really 27:58provided the problem with AI models, uh, yeah, I agree that they have, uh, 28:03you know, consumed the knowledge and they're not necessarily helping anyone. 28:05But I will say that most of the time people, when you get an 28:08answer from some AI model, you then wanna take an action going next. 28:11It's to your benefit to be able to send people back to sources that are trusted. 28:16Just like the way that right now, right? 28:17You, you search on Google, you search or whatever it is gonna send you 28:20the links that you're still probably gonna eventually wanna able to follow. 28:23Um, just like with social media, people are going to learn how to 28:27interact with this technology and they're going to learn to not just, 28:29you know, take the AI overview. 28:31And even though short term you might say, oh, who cares where this came from? 28:35Longer term we're gonna swing right back to, no, no, I care where this came from. 28:39So I want to be able to have that trust, have that trace, uh, the rest of it. 28:43So I, I think that actually if we 28:45start that work now, it's really gonna be helping ourselves in the future to 28:51set that up and, and have that going. 28:52Not in the least, because also would be nice to not kill Wikipedia. 28:56Please. 28:57We don have enough. 28:58I really depend on that. 29:00Um, so no, uh, again, I, I think we'll be able to, to get past this, but the more 29:05we can get, um, as Gabe was saying, actors from the larger companies that recognize 29:10this is actually in their own interests. 29:12Not just altruism. 29:13It really is in their own interest to do this. 29:16Um, the faster we get there the better. 29:17Yeah. 29:18And yeah, I think, uh, I really like that idea of kind of like, and maybe 29:20in the near term people are like, oh, the AI overviews just fine. 29:23I just use it. 29:24And then after a while people are like, Hmm, I don't know about that. 29:27And so like, there's a dip in traffic and then it kind of comes 29:29back as people are like, I gotta check the actual page or whatever. 29:31So 29:32I also think there's too much of a premium on recency to have all of 29:37this content just get baked into the model and never go back to the page. 29:41So something is visiting these pages to pull the moment of content and 29:46then feed that into the chat bot and return an answer which provides 29:49that pass through opportunity to go. 29:51Then click on the link, see the full source and everything else. 29:54So I agree with Marina. 29:55I don't think it's gonna, um, kind of necessarily revolutionize how 29:59the value, the value exchange, even if we have to maybe, you know. 30:02Continue to evolve a little bit. 30:05What I do wonder though, is like how can we move to more of a common crawl version 30:09of the world where, you know, these model providers, we didn't all crawl the 30:13entire internet ourselves independently. 30:17We all started from the common crawl snapshots of the internet and use that. 30:21And, you know, I think we do need, just like the community needs to come to these 30:26data providers, like for data providers like Wikipedia that are prioritizing 30:30the, you know, public dissemination of knowledge as part of their mission. 30:34We do need to work with them to set up more processes and offerings that are 30:39designed and tailored for model providers. 30:42So if, unless they're saying that don't crawl us, here's a robot's TXT, 30:46and we're not interested in these mo uh, data ever being used by models. 30:49And it doesn't sound like Wikipedia is saying that. 30:51Then, you know, it would be great to work together to identify, you know, here's a, 30:55a offering that Wikipedia is gonna start to more purposefully put together to reduce 31:00crawl traffic and improve the access of their information to large language models 31:05for this new mechanism of consumption. 31:08Yeah, and I think, I mean, that's why I'm kind of optimistic in some 31:10ways if some of this is, you know, almost Marina's model was like, 31:13just how easy is it to get the data? 31:15And it's like, well, if it's like a nicely. 31:18Produce data set which is updated and you know, refresh and great for your use. 31:22There's kind of no reason to write this reaper. 31:24And so there's a lot of need to kind of just like build these solutions that 31:27almost just like lower the cost of like, oh, we just access the data without 31:30having to hammer the servers all the time. 31:36Great. 31:37Well, uh, I'll move us on to our last, uh, story, which is mostly just kind of 31:40a fun one that popped up across my radar. 31:42Um, there was a story about the Beijing Humanoid Robot half marathon, which, 31:48uh, featured, uh, 1200 human runners alongside 20 robot teams from private 31:52companies and various state backed projects where they had kind of a robot 31:56running alongside the marathon runners. 31:58Um, and ends up being the humans are still good at this. 32:00The winner still was able to complete the ha ha half marathon, I think like an hour. 32:04Less than the humanoid robot, which still made it across the finish line, 32:07but at think two hours, 40 minutes and 27 seconds, you know, made it. 32:12Um, and I wanted to kind of bring this up both 'cause like the video is 32:15hilarious and you should check it out and it's just like a fun thing to watch. 32:18But also, 'cause I think we have been quite skeptical on this show about the 32:22entire craze around humanoid robots. 32:24I think every time I've brought it up as a topic, everybody has been 32:27like, this is never gonna be useful. 32:29This is just VC theater. 32:31I don't even know why people are talking about this. 32:33Um. 32:34But I don't know, there's kind of a part of me that kind of saw this happen is like 32:36this, like technology maybe as novelty as it is, seems to be getting quite good. 32:41Um, and so I guess may, Gabe, maybe you're new to the show, so 32:45maybe I'll kick it to you first. 32:46Um, are you similarly, like this is just an a play thing, or do you feel like 32:49you're like a little bit more bullish on. 32:51You know, humanoid robots being something that actually ends 32:53up being practically useful. 32:54Humanoid 32:55robot or not. 32:56I am a little bit bullish on the idea of really thinking wide about 33:01modalities and how AI as a general concept interacts with humans. 33:07Uh, you know, going back to what I said earlier, I think. 33:10Uh, the real novelty that was the jump between sort of the pre gen AI days and 33:15the gen AI days was taking the need for complex, uh, you know, output programming 33:21out of the picture and bringing the AI directly into a space where it 33:25could robustly interact with humans. 33:27Uh, and obviously we've, we've done a whole lot since then to. 33:30Uh, sort of blend those two things. 33:32You know, every AI model you hit behind a service is in fact 33:36a system and not just a model. 33:38But, um, I think the idea of extending that beyond the, the 33:43computer screen and the keyboard is actually really interesting. 33:46And I don't know necessarily whether chasing C-3PO is the right, 33:50The right direction for that. 33:52But I do think there are a lot of places where the physical interaction 33:57of robots, uh, is actually, you know, right now, you know, running a half 34:01marathon is a very constrained scope. 34:03And so part of me really wants to unbox what they actually built and understand, 34:07okay, now can that exact same robot go 34:10uh, also, 34:11to lean on a, an internet trope, fold my laundry for me. 34:14Right. 34:14Um, probably not, but it's really interesting to think about, uh, you 34:20know, the direction of extending into that physical modality as yet another 34:25place where AI can meet humans. 34:27So the, the concrete implementation here, I don't know, I'd have to read 34:31a lot of papers about it to understand whether there's value there, but the. 34:34The chase of bringing AI closer to where humans interact in more 34:39modalities I think is pretty cool. 34:40Marina, I feel like we're really trolling you this episode. 34:42Every single story, you're just like shaking your head 34:44and like muttering your breath. 34:46Do you wanna, do you wanna give your hot take on this? 34:49So, look, there's value in VC theater. 34:51If it means that VCs are gonna give money for actually valuable and you know, to the 34:56moment work in this direction, then great. 34:58The same people who built this robot know a lot about robotics in general. 35:01So they're gonna be doing a whole lot of work. 35:02So great. 35:03Bring on the theater. 35:04There's probably aspects here that are interesting in terms of artificial 35:07limbs, in terms of movement in general. 35:10I mean, you don't need this thing to be fast. 35:12You want it fast. 35:13Go get the MIT cheetah robot, that thing's gonna, run real fast, right? 35:17That's, that's not the point of this either. 35:18So honestly, hooray for theater and as long as it keeps attention 35:23on, uh, this and all the directions that Gabe just mentioned, these are 35:26things that we should continue to do. 35:28Just like basic scientific exploration. 35:30I feel like people have almost in the way that gen AI is right now 35:35and the speed at which things are going, everybody is just like, great. 35:38So what's the value? 35:39What's the value? 35:39What's the value? 35:40So now I'm gonna go ahead and disagree with myself 'cause this 35:42is what I say most of the time. 35:44We just, where, where's the value wise? 35:46There no value, but sometimes you need to allow people the time and 35:49the space and the money to do basic scientific research without really 35:53knowing what the hell the value is. 35:54Yet it will eventually come. 35:56Kate, I guess maybe I'll, I'll end this episode with a fun question 35:58for you is, are there other human robot competitions that you would 36:02wanna see robots competing in? 36:04Um, I don't know if there's particular use cases where you're like, I don't 36:07know about the science, but it would be really funny to watch X, Y, Z. 36:11Yeah, I, I, I don't know about that one. 36:13Folding laundry is certainly top of my use case list. 36:16I don't know about a competition there, but you know, I know I'm not interested 36:20in robots that can run faster than me. 36:22So just for many reasons, just not interested, not, 36:25I don't see a lot of value. 36:26And if the MO goal is to transport things faster, I think there's other 36:30modalities getting to Gabe's point about diversity of modalities that 36:34I think are gonna be prioritized. 36:35So, I, I agree with Marina. 36:37I think it's. 36:38Certainly, um, there is value to some of these demonstrations and setting 36:43targets that you then try and, and meet and exceed, but I do really wish that 36:48we could find, you know, non humanoid more fit for purpose, work on, on 36:53robots and prioritize some of that. 36:55I think just like we see with models where smaller, more fit for purpose models can 36:59drive a lot of value and you know, in many ways can be built more efficiently. 37:03I think we're gonna see the same, uh, in robotics more broadly and, 37:07you know, so I'm not super bullish on general purpose humanoid robots that can 37:12both run a half marathon and actually help me around the house in my day to day. 37:16Um, well that's a great note to end on. 37:18And I, I for one, as someone who's spending a lot of time folding 37:20tiny child laundry right now, I, I actually think that would be 37:22an incredible spectator sport. 37:24I would be very excited about the, uh, humanoid robot laundry folding ESPN four. 37:304:00 AM in the morning, uh, televised competition. 37:33Um, well that's all the time we, we have, uh, for today. 37:36Uh, Kate, Marina,, as always, great having you on the show. 37:38You're a dynamic duo every time you come on. 37:40I feel like, I just, like, there's all these comments where I'm like, oh yeah. 37:42It's like I never really thought about it that way. 37:44Um, and Gabe, welcome to the show for the first time, hopefully we'll have 37:47you on, uh, at some point in the future. 37:49Thanks to all you listeners, if you enjoyed what you heard. 37:51You can get us on Apple Podcasts, Spotify, and podcast platforms 37:54everywhere, and we will see you next week on Mixture of Experts.