Learning Library

← Back to Library

ChatGPT 5.2: Agentic Future Delegation

Key Points

  • GPT‑5.2 is a fundamentally new, “agentic by default” model that can autonomously process massive datasets (e.g., 10 000 rows), perform analyses, and generate finished deliverables like PowerPoints, docs, and Excel files with reliable accuracy.
  • The breakthrough lies not just in speed but in the ability to compress work that would normally take six‑to‑eight hours into a 20‑minute run, dramatically reshaping productivity expectations.
  • To harness this power, users must master a new skill: precisely defining the scope and desired outputs of a task so the long‑running agent can be delegated work effectively.
  • Despite its advanced capabilities, the model cannot replace an entire job; clear, scoped instructions are still essential for meaningful results.
  • The improved quality of generated artifacts (e.g., usable PowerPoint slides) marks a leap from earlier versions and signals the kind of competence expected of professionals by 2026.

Full Transcript

# ChatGPT 5.2: Agentic Future Delegation **Source:** [https://www.youtube.com/watch?v=821UqXHineU](https://www.youtube.com/watch?v=821UqXHineU) **Duration:** 00:14:38 ## Summary - GPT‑5.2 is a fundamentally new, “agentic by default” model that can autonomously process massive datasets (e.g., 10 000 rows), perform analyses, and generate finished deliverables like PowerPoints, docs, and Excel files with reliable accuracy. - The breakthrough lies not just in speed but in the ability to compress work that would normally take six‑to‑eight hours into a 20‑minute run, dramatically reshaping productivity expectations. - To harness this power, users must master a new skill: precisely defining the scope and desired outputs of a task so the long‑running agent can be delegated work effectively. - Despite its advanced capabilities, the model cannot replace an entire job; clear, scoped instructions are still essential for meaningful results. - The improved quality of generated artifacts (e.g., usable PowerPoint slides) marks a leap from earlier versions and signals the kind of competence expected of professionals by 2026. ## Sections - [00:00:00](https://www.youtube.com/watch?v=821UqXHineU&t=0s) **ChatGPT 5.2 – Agentic Data Mastery** - The speaker highlights the upcoming ChatGPT 5.2 as a genuinely agentic model that can autonomously process huge datasets, produce polished PowerPoints, Excel reports, and executive narratives, emphasizing the new skill of delegating work to such an AI. - [00:03:37](https://www.youtube.com/watch?v=821UqXHineU&t=217s) **Effective Prompt Framing for Advanced Models** - The speaker stresses that with newer, more capable, agentic models like Chad GPT 5.2, users must clearly define the task and scope—otherwise the model will guess—making precise problem framing a universal skill, illustrated by comparative tests against Gemini 3, Claude Opus 4.5, and ChatGPT 5.1. - [00:08:24](https://www.youtube.com/watch?v=821UqXHineU&t=504s) **Leveraging Large Context for AI Tasks** - The speaker emphasizes that the key advantage of advanced models like ChatGPT 5.2 is their ability to process extensive, varied datasets within a large context window, enabling them to solve complex, data‑intensive problems such as customer service analytics and multi‑source analysis. - [00:11:58](https://www.youtube.com/watch?v=821UqXHineU&t=718s) **Leveraging GPT‑5.2 for Agentic Workflows** - The speaker outlines how to intentionally steer the upcoming 5.2 model, allow it extended processing time, and use it as a flexible, agentic executor for deep analyses such as P&L reviews, acquisition assessments, and personal budgeting, while emphasizing the need to choose the right problems and provide appropriate data. ## Full Transcript
0:00Chad GPT 5.2 time traveled back to see 0:04us here. I am convinced that this is a 0:06model that shows us what the future is 0:08like for 2026. It's not an incremental 0:11upgrade, guys. I know it's positioned 0:13that way, but it's actually got some 0:16capabilities that I haven't seen in 0:17other models that I want to lay out here 0:19so that you understand what they are and 0:21you can figure out for yourself whether 0:23the model is right for you. First and 0:25foremost, this model is agentic by 0:29defaults. So if you think about models 0:32on a range of how long they can run and 0:34execute tasks, this is the first 0:37generally available model where it's 0:40very very easy to get it to do a 0:43tremendous amount of work on a huge 0:46bucket of inputs like a data set with 0:48thousands of rows. I tried it with a 0:50data set with 10,000 rows, right? It can 0:52do all of that, compute against it, 0:55develop insights, come back with a 0:57PowerPoint, come back with a doc, come 0:59back with an Excel spreadsheet, and it 1:01actually works. That means that it's 1:04accurate. It's coherent. It's cogent. 1:07It's thoughtful. It's able to craft an 1:09executive narrative. Guys, the 1:11PowerPoint is not nearly as gnarly as it 1:14was in 5.1 and 5.0. The PowerPoint 1:16artifacts actually work now. It's 1:19wonderful. But this creates a skill 1:21problem for us, doesn't it? Because what 1:23we have to do is we have to figure out 1:27how do we now define work that is ready 1:31to be delegated for that period of time. 1:33And that's a new skill for a lot of us. 1:36For many of us, we have been trying to 1:38figure out how to make these models help 1:41us do our work faster all year long. And 1:44that's been most of the conversation 1:45I've had with folks. Guess what? the 1:48models keep getting better and we have 1:50to keep scaling up. And in this 1:52situation, the skill that we need to 1:54learn whether we're technical or 1:55non-technical is how do we define a 1:58piece of work correctly so that we can 2:02assign it to a longunning agent. That is 2:04what feels like 2026 about chat GPT 5.2. 2:09That's what feels novel, new, and super 2:11interesting. And if you can't define 2:14that work, you are going to be behind 2:16people who can really define it well and 2:19come out with a fullyfledged analysis 2:21from a deep data set or a deep problem 2:23in the code or what have you and then 2:26get an answer that they can use and run 2:28with that would normally have taken them 2:30hours. Because when I take and I'm not 2:32kidding 20 minutes, 30 minutes, 40 2:33minutes on a chat GPT 5.2 two task which 2:35I did today. It's it's really good and 2:39it's better than work that would have 2:40taken me four or five hours to do. And 2:42so it's not just about can it save me 20 2:45minutes. It is understanding that the 2:47model can do in 20 minutes what would 2:49have taken someone six or eight hours to 2:51do and how do you understand that block 2:54of work and give it to the model. Now 2:57you might think if it can do six or 2:58eight hours of work can it just do my 3:00job? The answer is no. It needs clear 3:04scope. When I talk about the skill to 3:07delegate to the model, the first thing 3:08to do is to be able to define what 3:11output you want. A scoped output that 3:14matters. If you want a PowerPoint deck, 3:15it can do that. You have to define what 3:18you want there. If you want a word doc, 3:19it can do that. If you want an Excel, it 3:21can do that. Specify. Be clear about 3:24what you want. You also need to be 3:26really clear about what you need from 3:29the inputs. Especially if you're going 3:31to use that nice big context window and 3:33you're going to put a bunch of stuff in. 3:35Please explain to the model what is in 3:37the box and what you want the model to 3:40do with it. Because if you don't, the 3:42model's going to fill in its best guess 3:44and it's going to try and make it 3:46intelligible as best it can and you may 3:48or may not get what you want. And that 3:50has higher stakes now, doesn't it? One 3:52of the big things that shifted in the 3:54last 6 months is that we are no longer 3:57in a world where instant responses are 3:59the best a model can do. The best a 4:01model can do is often longer running. 4:03And so if you're in a world where the 4:04model can take a while to come back with 4:07a response, you better get it right. 4:09You'd better be correct in your problem 4:11framing. And that's not just an 4:13executive skill set anymore. That's an 4:15everybody's skill set. All of us need to 4:17learn more about framing problems and 4:20chunking problems into scopes of work 4:22that that can fit with a model that is 4:26truly agentic. And the reason I'm 4:28emphasizing that here is because Chad 4:30GPT 5.2 is so widely distributed. 4:32Everybody's going to get it because 4:34everybody has Chad GPT. So, we all need 4:36to learn this. Now, now you might be 4:38wondering, how does this compare to some 4:40of the other models out there? Well, I 4:41want to give you some very specific 4:43comparison notes that I've been seeing 4:45in early testing because I did a cross 4:47analysis where I gave the same 4:48assignment to different models to see 4:50what the quality would look like. I 4:51tested against Gemini 3. I tested 4:54against Claude Opus 4.5. I tested on 4:56Chat GPT 5.1 as well just to see what 4:59the sense of of of a difference is 5:01versus 5.2. I think that I'm getting a 5:04real clear sense of where these 5:05different models stack up. One of the 5:07things that is standing out to me is 5:09that the ergonomics of the model matter 5:12a lot. By ergonomics, I mean how do you 5:14have the full environment around the 5:17model feel comfortable like a good 5:19ergonomic chair so you can use it for 5:22useful work. That's not just comfort. 5:26That's actually value. Specifically, 5:28Gemini 3 has really poor user ergonomics 5:32right now. They have embedded Gemini 3 5:34inside Google products and you can 5:37access Gemini 3 in the developer studio 5:40and you can access Gemini in the mobile 5:43app. But in none of those places is it 5:46easy to throw a bunch of data to throw a 5:50bunch of docs into the model and say 5:52please come out with a fully finished 5:54output. That is just not the product 5:56that Google has built. And so even if 5:59the brain power is there to do 6:01meaningful work against these artifacts 6:04and analyze it and come back with a 6:06fully featured output, you can't get to 6:09it. I could not upload a PowerPoint to 6:11Gemini 3. I could not upload a 6:13PowerPoint to or an Excel or a CSV. It's 6:17just not good, guys. you you you have to 6:20have the ability to put a lot of data in 6:22if you want to do complex work and it's 6:24a problem if you can't do that. And so I 6:28love Gemini 3. I did a great review on 6:31it. I still use it. I love their image 6:33generator. It's a smart model. I use it 6:36for thinking a fair bit. But the 6:38ergonomics are a and they really pop out 6:41when you compare it to Chat GPT 5.2 6:44because Chad GPT 5.2 YouTube will take 6:46anything. Like you can throw anything in 6:48there. It will take it all and it will 6:50just chew on it. You can throw a 6:52screenshot and a CSV and a doc and a 6:55PowerPoint and it will just chew it all 6:57and process it and come out with 6:59something useful. And I think that's 7:00really really helpful. And I think that 7:02one of the things that really popped as 7:04a difference in my test between 5.2 7:07is that the ability to intelligently 7:10coherently with less hallucinations 7:13process this data is way up. And that 7:15showed up in their benchmarks. They saw 7:16like 38% less hallucinations or 7:19something like that. And and it just it 7:21pops like you can see it. You can you 7:23can see the coherence. Now comparing it 7:25to Opus 4.5 is interesting because the 7:27ergonomics in Opus 4.5 are also quite 7:29solid. You can throw in a wide variety 7:31of input documents. I like the way Opus 7:344.5 is able to craft effective output 7:37artifacts just like Chad GPT 5.2. And so 7:40if I were to look for a difference 7:42between the two, I think the thing that 7:44I want to call out is first the way the 7:46models are architected is very very 7:48different. Chad GPT 5.2 especially in 7:52thinking mode which is a very different 7:53mode. If you're using instant mode, it's 7:55not the same thing. Chad GPT 5.2 7:58thinking mode is a longunning thoughtful 8:02intentional model. It takes a while to 8:04respond. It does very thorough work and 8:06these days it now does artifacts well 8:09too. It really does. there's not really 8:10that gap on PowerPoint functionally 8:12speaking. Opus uses tools instead of 8:15reasoning. And so Opus will work for a 8:17while, but it's using tools as a 8:19non-reasoning model to get that work 8:20done. So it's a very different approach. 8:22I like the aesthetics of the PowerPoint 8:24that Opus 4.5 produces slightly better. 8:28The functionality is about the same. 8:30Like from a functional PowerPoint 8:31narrative perspective, it's about the 8:32same. And critically, the thing that 8:35gives Chat GPT 5.2 to an edge is that it 8:39can take so much data to solve your 8:42problems. And that's why I started this 8:44conversation saying pay attention to how 8:47long these agents can work. Because if 8:50you were going to give an agent a 8:51meaningful task, that only really works 8:54if you trust it with a ton of data. If 8:58you give it a lot of data to work with 9:00and ask it to handle a complex task. 9:03Otherwise, even in thinking mode, it 9:05won't take that long. and you won't have 9:07solved that meaningful a problem. And so 9:10I think the thing that we need to shift 9:12toward is a world where we recognize 9:16that increasingly the models have a 9:20better understanding across larger 9:22swaths of data than we do. So maybe it's 9:25a a customer service set of tickets that 9:28we need to analyze. Maybe it's hundreds 9:31of Twitter responses from a question 9:33that we had. Maybe it's uh a bunch of 9:35Stripe transaction data. Maybe it's a 9:38big Excel spreadsheet of customer 9:40issues. You get the idea, right? It's 9:42anything that has that sort of very 9:45large variagated data types all in one 9:48big place, right? Because you could have 9:50like customer tickets in one hand, you 9:52could have transcripts and recording in 9:54another. The data can be quite 9:55variegated. It's a big enough context 9:57window you can throw it all in there and 9:59you can ask it to make sense of it. And 10:01it does. and it's able to translate it 10:04into something useful. I think this is a 10:06little bit of an intangible, but one of 10:08the things that comes out when you have 10:10a model that is strong at coherence that 10:12reduces on hallucinations that has the 10:14tools to build something like a 10:16PowerPoint well is the ability to build 10:19narrative comes as an emergent property. 10:22And so what I noticed is it's able to 10:24take data that I don't necessarily have 10:26a clear story for and it's able to pull 10:28it in and say there's a story here. 10:31here's the overall story and here's why 10:33I know that and you can check it and 10:35prove it because of course you do. You 10:37have to go in and check and and see that 10:38it actually works. And so if you're 10:41looking at 2026 and you're asking 10:43yourself, what are the skills I need to 10:45thrive? How do I build this into my 10:47teams? I would say the number one skill 10:50that you're going to need in chat GPT 10:525.2 too and in the other models that 10:54follow not just from Chad Chad GBT but 10:57from Gemini from XAI from Anthropic 11:01you're going to see more agentic models 11:03and your number one skill needs to be 11:06grow my ability to delegate we are 11:09moving from a world where execution with 11:12models was the story for 2025 delegation 11:15to models is going to be the story of 11:182026 11:20we're not ready we're not ready. We're 11:23not ready with the data side. We're not 11:25ready with the skill side. We don't know 11:27how to frame problems. Look, I the first 11:29thing I did when I started getting into 11:315.2 and seeing what it could do is I 11:34went over to 5.2 and I asked it to start 11:37to help me think through prompting this 11:40model differently because we have to 11:42think about prompting not as give me a 11:45response now, but as let me give you a 11:48lot of stuff and then go away and think 11:51about it. Now, eventually we're going to 11:52get to a world where I think we have 11:54more interaction patterns with running 11:55agents and you can interrupt the agent. 11:57We're starting to see hints of that. 11:58We'll see more of that in 2026. But the 12:01skill for now is really intentionally 12:03aim the model in the direction you want 12:05it to go and then focus and make sure 12:08you have the right stuff and then give 12:10it give it time to work. Let it work for 12:12a while. It is not unusual to see a 12:14model like 5.2 two work for 20, 30, 40 12:18minutes and it's not like deep research 12:19because deep research comes back and it 12:22just gives you a web report and it's 12:23very well written. It can be 50 pages. 12:255.2 thinking will come back in a similar 12:28amount of time but it will give you much 12:29more control over what you get. You can 12:31define the output type that you want. 12:33You can define the kind of analysis you 12:35want. It's like a much broader Swiss 12:37Army knife versus the scalpel that is 12:39deep research. So if you're wondering 12:41where to put this into your workflow, I 12:43would say 5.2 to thinking is a agentic 12:49workflow executor that is almost more 12:52powerful than we're ready for. It is 12:54something that if you know how to 12:56delegate well, it is going to eat work 12:59for you. You want to analyze a P&L, let 13:02it let it analyze the P&L for you. Let 13:04it take the first pass. You want to 13:05analyze an acquisition, let it do that. 13:07You want to analyze your investments or 13:08your personal savings and budget, let it 13:11do that. This thing loves to solve 13:14problems. And so really the rate limiter 13:17for us, the question for us is do we 13:19have the taste to find the right 13:22problems to solve? Do can we can we 13:23locate the data for it? Can we throw the 13:26data in and then can we give it clear 13:29enough directions about the output and 13:31the kind of analysis it needs to run in 13:33order to get successful outcomes because 13:35the stakes are higher. Now if you're 13:37running Chad GBT 5.2 too for 20, 30, 40 13:40minutes and you didn't give it the right 13:42directions. Your feedback loop is slow. 13:44You're going to be like, "Oh no, now I 13:45have to redo it. It's going to be like 13:47another hour out of my day to get this 13:48done." So, our prompting skills are now 13:50higher leverage because it's so 13:53important. And so, I put together some 13:55prompts to make sure that we have a good 13:57sense of what this looks like. But 13:59beyond prompting, the key thing that I 14:01want to call out is we need the soft 14:04skills to delegate better, to understand 14:06those problem frames. And that's what I 14:08want to leave you with because I believe 14:10that that is the key skill for 2026. And 14:13I think that is what 5.2 shows us in a 14:15way that no other model does. It will 14:18eat entire workflows because it is so 14:21good at correct coherent longunning 14:24agentic execution. I think they kind of 14:26undersold it as a 0.1 upgrade. I think 14:29it's bigger than that. But you tell me. 14:31You test it out. You tell me what you 14:33think. I'm really curious. I love the 14:35model. It's going to be a lot of fun to 14:36use.