Learning Library

← Back to Library

AI 2026: From Hype to Results

Key Points

  • I’m optimistic for 2026 because AI will finally be judged on whether it works in real‑world applications rather than on flashy demos or benchmark scores.
  • The hype bubble burst in 2025 (e.g., a disappointing ChatGPT‑5), prompting conversations to focus on edge‑case, multi‑agent, and tool‑use systems that actually ship.
  • A torrent of new capabilities—Claude Code, reasoning models, Codeex, Nano Banana/Pro—appeared in 2025, giving us a “4K” view of what practical AI systems can achieve.
  • My hope centers on the surrounding ecosystem, especially talent that can blend protocol design, interface engineering, verification loops, and customer focus into a single, cohesive role.
  • While we still need to build organizational structures to support these roles, such interdisciplinary experts are emerging and will be crucial for delivering AI‑driven value in the coming year.

Full Transcript

# AI 2026: From Hype to Results **Source:** [https://www.youtube.com/watch?v=RVviMEfaJUY](https://www.youtube.com/watch?v=RVviMEfaJUY) **Duration:** 00:15:40 ## Summary - I’m optimistic for 2026 because AI will finally be judged on whether it works in real‑world applications rather than on flashy demos or benchmark scores. - The hype bubble burst in 2025 (e.g., a disappointing ChatGPT‑5), prompting conversations to focus on edge‑case, multi‑agent, and tool‑use systems that actually ship. - A torrent of new capabilities—Claude Code, reasoning models, Codeex, Nano Banana/Pro—appeared in 2025, giving us a “4K” view of what practical AI systems can achieve. - My hope centers on the surrounding ecosystem, especially talent that can blend protocol design, interface engineering, verification loops, and customer focus into a single, cohesive role. - While we still need to build organizational structures to support these roles, such interdisciplinary experts are emerging and will be crucial for delivering AI‑driven value in the coming year. ## Sections - [00:00:00](https://www.youtube.com/watch?v=RVviMEfaJUY&t=0s) **Optimism Shifts to Practical AI** - The speaker is hopeful for 2026, marking a transition from hype‑driven benchmarks to real‑world, results‑focused AI applications following a 2025 disappointment in consumer expectations. - [00:05:11](https://www.youtube.com/watch?v=RVviMEfaJUY&t=311s) **Designing Agentic AI Workflows** - The speaker argues that by embedding validation rules, graceful degradation, and deterministic code for routing, counting, and retries, we can slot LLMs into narrowly scoped, high‑value steps within disciplined workflows, turning them into robust production‑grade AI‑native experiences that go far beyond simple chat. - [00:11:01](https://www.youtube.com/watch?v=RVviMEfaJUY&t=661s) **Future of Graphical AI & Dual Fluency** - The speaker explains that multi‑agent generative workflows will enhance retention, foresees image‑driven AI tools becoming commonplace by 2026 after breakthroughs such as Nano Banana Pro, and predicts that the job market will reward professionals who master both AI behavior and the core craft of their role. - [00:14:25](https://www.youtube.com/watch?v=RVviMEfaJUY&t=865s) **Future of OTA-Updated Home Robots** - The speaker predicts a 2026 boom in household robots, stressing that market leaders will be those who can reliably ship devices and provide regular over‑the‑air software updates to keep their robot “brains” continually advancing. ## Full Transcript
0:00So, I've been spending time thinking 0:01this holiday season about what I'm 0:03optimistic for for artificial 0:05intelligence and all of us in the year 0:07ahead. And I think it comes down to 0:08this. I'm optimistic for 2026 and AI 0:12because we are exiting the era when AI 0:15is going to be judged by how clever the 0:18release is, how fancy the benchmark is, 0:20how exciting the demo is, and we are 0:23entering the era where it's going to be 0:25judged by whether it works. And I love 0:27that because that means we're actually 0:29getting to a point this year where we 0:31can focus on delivering results with AI. 0:35And that's hard, but it's meaningful 0:37work. And I think that there is really 0:40like the bubble of hype really burst in 0:422025. Felt it when chat GPT5 was 0:45disappointing to so many consumers. And 0:48I think the most instruction instructive 0:50conversations I've had over the second 0:52half of the year especially, they've not 0:55focused really on model road maps. 0:56They've not focused on benchmark charts. 0:59They've been about the critical edge 1:03casriven work that shows up when you try 1:05and ship real systems, real multi- aent 1:08systems, real tool use systems, real 1:12systems that enable a human to do much, 1:16much more than they could do before. And 1:18so I feel like now as we enter the new 1:20year, we're getting to a point where you 1:24can actually start to imagine the 1:29details needed to get an intelligence 1:32layer that we can all benefit from in 1:34the new year. Something that helps our 1:36work to go farther. And for a lot of 1:382025, we were coloring in those gaps 1:40with hope because we couldn't imagine 1:42it. Like if you think back over the 1:45year, Claude code is less than a year 1:47old. It was out in private beta in 1:49February. Uh we had just had reasoning 1:52models at the start of the year in 2025 1:54and they were very new. These things are 1:56moving really quickly. Codeex didn't 1:58exist until partway through 2025. Nano 2:00Banana and Nanobanana Pro both came out 2:03in 2025. And so all of these things that 2:06feel like essentials for the new systems 2:08in 2026 came into being over the course 2:11of the year and enabled us to start, I 2:14guess I would call it seeing in 4K, 2:16right? We're starting to see in high 2:17definition what's possible with these 2:19models in a way that we had to guess at 2:22before. And so that's why a lot of my 2:24optimism for the new year is about the 2:28ecosystem around AI and not just about 2:31AI itself. And I think the optimism I 2:34have is also about the talent that goes 2:36with that ecosystem. I think one of the 2:38things I'm really excited to see is 2:41talent that can hold protocols and 2:44interfaces, technical details, 2:46verification loops, and the passion for 2:48the customer together so that the 2:50technical reality and the job to be done 2:52sit in one head, not in two or three or 2:55four or five heads at a time. And I 2:56think we're getting closer to those 2:58kinds of roles. We definitely have more 3:00development work to do on our people org 3:02side so that more of those roles are 3:04published and available etc. But I see 3:08people who can do that emerging more and 3:11more and they're incredibly valuable 3:13wherever they operate. So with that, let 3:15me get a little bit more detailed in the 3:17spirit of the season and talk about some 3:19bets that I feel optimistic about as we 3:23head toward 2026. One that I think is 3:25really interesting that we don't talk 3:27about a lot is that I feel optimistic 3:29that our protocols and our process are 3:32going to start to matter even more than 3:34our prompting. And so we've been 3:36treating prompting, we've been tempted 3:38to treat prompting as a very primary 3:40interface. And that was true in the chat 3:43era. And now I think we're going to 3:44start to treat it more as a layer in a 3:47more standardized tool chain for agentic 3:49workflows. And so the teams that win 3:52won't be the ones that necessarily have 3:54the cleverest instructions. They'll be 3:56the ones where the systems can reliably 3:58call the tools and pass the structured 4:00outputs and hand off work between 4:01components and where they can reliably 4:03recover when something goes wrong. That 4:06means that 2026, what I'm hopeful for is 4:08that we will be reinventing the wheel 4:10less. There'll be less bespoke glue 4:12holding everything together and more 4:14composable systems. Another thing I'm 4:16optimistic for is the idea that we will 4:20take constraints seriously in AI. That 4:23sounds like a funny thing to be 4:24optimistic for, but I I think it 4:26matters, right? Because the constraints 4:28are the difference between content and 4:31software. If you're just saying, "Write 4:32me 200 words or write me a story about X 4:35or Y or help me with this prompt," 4:38you're really unconstrained and you're 4:40just asking for a chat response. But as 4:42we move more into agentic workflows, 4:44we're going to be giving our LLMs very 4:47tight constraints in order to enable 4:50them to do useful, repeatable work at 4:52scale. And that's why I'm saying like I 4:54think we're moving through this 4:56transition where we're going from LLMs 4:58as content generators to LLMs as 5:01software. And that's a really cool 5:03journey to see. And I think a lot of 5:05teams that start to take constraints 5:06seriously are going to get the layouts. 5:10They're going to get the validation 5:11rules. They'll get the graceful 5:12degradation, the repair steps, the 5:14fallbacks, all of that baked in. And 5:16before they know it, their workflows are 5:17going to be in a spot where you can 5:19actually call it working software in 5:21production. And that's going to enable a 5:23new class of AI native experiences that 5:26go way beyond chat. And we really have 5:28all the building blocks for that. and 5:29the only thing standing in the way is 5:31just the discipline to start to take 5:33these LLMs and slot them in correctly. 5:35Another one that I'm excited about is 5:38really getting agentic workflows that 5:40understand where AI goes in those 5:42workflows. I think we've spent a lot of 5:442025 5:46thinking that LLMs could do everything 5:49in the workflow. And I think where we're 5:51coming to at the end of the year is that 5:53more and more LLMs are useful for very 5:56high value roles that are narrowly 5:59scoped within agentic workflows that 6:01have very specific deterministic 6:03transforms and checks associated with 6:05them, very specific tool calls. Really, 6:08that's all about deciding and defining 6:10where that model is good at generating 6:12smart tokens and abstracting everything 6:15else away in the workflow so it doesn't 6:16have to do that. So, we let the code do 6:18what the code's good at. We let it 6:20count. We let it route. We let it 6:22validate. We let it retry. We let it 6:24diff. We don't ask the LLM to do that in 6:27the prompt. And some people would say 6:29that's anti-agent, but to me, that's 6:31very pro- agent. It's actually 6:33understanding what LLMs are good at and 6:35starting to build systems where they 6:36thrive. It's pro- reliability. So, I'm 6:39really excited to see teams start to 6:41pick that up. Another one that I'm 6:42really interested in, this is going to 6:43sound theoretical, but we're going to 6:45get practical here. I'm excited that 6:47teams are understanding how entropy 6:50works with LLM systems. Uh I think in 6:532025 a lot of teams accidentally built 6:57systems that increase entropy and chaos. 7:00They had too many unconstrained steps, 7:02too many loops, too many opportunities 7:04for the model to get creative in the 7:06wrong place. And in 2026, I think those 7:09same builders are going to be the ones 7:12who start to understand that LLMs don't 7:15have to be drivers of entropy. People 7:18sometimes look at these token generators 7:19and say they're just uncontrolled. 7:21They're probabilistic. You can't manage 7:22them. And one approach, which I talked 7:24about earlier in this video, is to say, 7:26well, let's put some business rules 7:27around it. But I actually think a higher 7:29level approach, which is sort of what 7:30I'm getting at here, is to look at LLMs 7:33as potentially entropy reducers or 7:36decreasers. If you can actually 7:38structure where the LLM lives against 7:42your business outcomes correctly, then 7:45what was magical before can be a kind of 7:48disciplined magic now. And I think we're 7:50starting to see that in the chat driven 7:52experiences we have off of chat GPT, off 7:55of Claude, in product. I think we're 7:57starting to see that in some of the AI 7:59native interfaces. TL Draw comes to 8:01mind. That's definitely one that feels 8:03like magic but is actually extremely 8:05structured. Another one is the way Figma 8:07is handling AI at the end of 2025. 8:09Capsules is a good example. These are 8:11all places where LLMs are being 8:15harnessed in ways that produce more 8:18compelling and coherent and beautifully 8:20designed experiences that on the on the 8:23whole decrease entropy. It is there's 8:26less entropy in the system when I can 8:28get the answer I need inside the 8:30interface I have and I don't have to 8:31spray tokens everywhere finding some 8:33answer that I'm looking for on the 8:35internet as a whole. There's less 8:37entropy in the system when I can talk to 8:39my Figman design and get that correctly 8:41laid out and then get it directly into 8:43cloud code. And so entropy is a very 8:46high order way of talking about what 8:48we're doing when we design agentic 8:49systems. And I think teams are starting 8:51to recognize that you can design systems 8:54that are high entropy or low entropy 8:57depending on where you harness and how 9:00you harness the LLM against a larger 9:02customer outcome. And so my 9:04encouragement, the thing I'm excited 9:05about is that teams are starting to 9:06intuitively grasp this even if they 9:08don't have the language. And that means 9:10that they are starting to recognize that 9:13LLMs need a lot of harnessing to produce 9:16beautiful experiences. But you can do 9:18that. And if you do do that, you can 9:20deliver things that are way beyond what 9:22chat GPT brings you. And that brings me 9:24to another area where I'm optimistic. I 9:27think we are just at the beginning of a 9:30post chat GPT software future. I think 9:33that one of the things I'm truly excited 9:35about is that cursor has shown that even 9:38if you are quote unquote a rapper, you 9:40can absolutely thrive in the middleware 9:43layer. And that's a really interesting 9:44insight coming out of the year. And I 9:46think there's a lot of room to run, 9:50especially in non-technical areas for 9:52middleware in 2026. And a lot of it 9:55comes down to what I've been talking 9:56about with designing good agentic 9:57systems, decreasing entropy, making it 9:59more beautiful and useful to the 10:01customer. And you know, to be honest, 10:03one of the things that I think is really 10:05critical for that that we also are 10:08starting to learn is figuring out how to 10:10answer requests as if they're not all 10:12the same. You know, chat GPT trained us 10:15to answer requests as if they're all the 10:17same. But one of the characteristics of 10:19these new systems is they recognize that 10:21users have really different needs and 10:22you can build different experiences 10:24around them. Like if we talk about 10:26generative UI, generative UI is really 10:28downstream of the core insight that you 10:31can route users to experiences that 10:34matter to them outside the chatbot in 10:36ways that are beautiful and useful. If I 10:38want to cancel my phone bill, I should 10:41be able to just get a generative UI 10:42pulled up and do that. I shouldn't have 10:44to go six clicks deep. And that's we're 10:46just at the beginning of figuring out 10:48how to map the customer intent into 10:51probably a power law distribution of 10:53user utterances so that we can start to 10:55say so you know 90% of my user 10:57utterances are very common, very usual. 10:59This is how I handle them. But then I 11:01use like a great multi- aent workflow 11:03and generative UI to handle that long 11:04tail and suddenly it becomes a really 11:06powerful experience and it acts to drive 11:09retention to drive engagement across the 11:12entire population. Another area where 11:14I'm really optimistic is uh what I would 11:16call sort of the graphical AI world is 11:19going to become really normal in 2026. I 11:22think this is a downstream breakthrough 11:24of Nano Banana Pro. We're going to see a 11:26lot more work product that is just 11:28generated entirely as artifacts rather 11:30than pros. Like one of the very specific 11:32implications I think this has is that we 11:35will see just slideware that's very 11:38normally just images now because it's so 11:40easy to edit and regenerate images. You 11:43can already edit nano banana images 11:45inside manis and just regenerate and 11:47it's very trivial to get a new deck. And 11:49so when we live in that world where 11:51images are essentially solved, I think 11:53that opens up for us a lot of really 11:56interesting build opportunities in the 11:58new year around imagerriven AI. And 12:00we're just beginning to scratch the 12:02surface with that, but I'm really 12:03excited about. I think another one 12:06that's really interesting to me is 12:08careers are really repricing around dual 12:11fluency right now. So the market is 12:13going to start to reward people who can 12:15do two things at once. One is understand 12:17how AI behaves at a high level of detail 12:20and two is understanding the underlying 12:22craft of their role and the customer. 12:25And most organizations are still split 12:27right now between like an AI person and 12:29then like a domain person that AI person 12:31pairs with. I am wondering if in 2026 12:34we're going to start to see more roles 12:36that sort of put them in the same head 12:38because if you try and pair an AI 12:41person, even a very technical AI person 12:43with a domain person, the head has only 12:45half the answers. And I think that 12:47companies that can find those fully 12:49rounded people who understand a 12:51particular domain well and who also 12:53understand how a AI behaves in high 12:56fidelity, they are going to be highly 12:58sought after. And we're going to start 13:00to see HR systems rewrite jobs to get 13:03those people because people are starting 13:05to recognize the value and the alpha in 13:07the market and they have a year under 13:09their belts with AI and they're now 13:10training themselves and able to build 13:12things that they weren't able to build 13:14before and show their talent in a way 13:16that's really useful. I think the last 13:17thing I want to call out that I'm 13:18optimistic for is that I think robotics 13:21is going to have a huge year in 2026. Uh 13:25I'm not really talking about humanoids 13:26only. I'm talking about robotics more 13:28broadly. I think we have had a year 13:31where we started to put in a lot of 13:33groundwork on reinforcement learning. I 13:35don't know if you recall, but back in 13:37January of 2025, 13:40Nvidia announced their digital 13:42warehousing concept and this idea that 13:44you would give robots digital thousands 13:47of digital years of experience in 13:49simulated warehousing environments so 13:50that they would be safer in real 13:52warehousing environments. Imagine that. 13:54We've had a year to run on that. Toward 13:56the end of this year, we had a 13:59breakthrough where we're now able to use 14:01personal POV cameras looking at hands to 14:04allow robots to infer how hands move and 14:07learn from human hand movements. The the 14:10arc of the year is really around getting 14:12our learning in order 14:14so that in 2026 we can start to rapidly 14:19scale out LLMdriven robotic capability. 14:23It's going to look like constrained 14:25environments at first. It's going to 14:27look like cheaper compute at first for 14:30deployment in designated areas of 14:31warehouses. There is absolutely going to 14:34be a big push on home robotics in 2026. 14:38I don't know if that means we'll finally 14:39get the home robot laundry machine. We 14:41will see. But to my mind, I think what 14:43I'm most interested in is that the 14:46winners in this space are going to be 14:49the ones that have the ability to 14:51reliably ship and update the brains of 14:54the robots they're shipping so that 14:56consumers who are used to seeing these 14:58LLM updates every 2 or 3 months don't 15:01feel left behind when their household 15:03robot is shipped to them in November and 15:05there's a new software drop in January. 15:07I think that we're going to see 15:09essentially ecosystems start to develop 15:11where people will say the robot 15:12primitives are all there. Uh and people 15:15could be business owners, could be 15:17humans, uh who own robots at home, 15:18whatever it is, but I want overtheair 15:20updates that ensure that the robot's 15:22brain keeps getting smarter and it can 15:24use those fingers or it can use the 15:25pinchers or whatever the robot has more 15:27and more effectively over time. And I 15:29think that that's one of the pieces that 15:31we have all the building blocks for and 15:32I'm sort of optimistic to get there in 15:342026. What are you optimistic for in the 15:37spirit of the holiday season for 2020s?