
2025 was a year that was saturated in AI news, from Deep Seek, through claims of economic “bloodbaths,” to GPT-5, Sora, and Chatbot girlfriends. Frankly, it was exhausting. As we now look back on 2025 an interesting question arises: all in all, did this end up being a good or bad year for AI? To help me answer this question, I’m joined by hard-hitting AI commentator Ed Zitron, who's been everywhere in the media in recent months helping to make sense of the wild claims being thrown in the public’s direction. Together we go through the biggest AI stories of the year to try to make sense of what just happened.
Loading summary
A
So much happened in the world of AI in 2025 that it can actually be hard to keep track of it all. I mean, remember Deep Seek? That was in 2025, as was Dario Amade saying that we were going to lose half of white collar jobs to AI as well as GPT5's release. The release of Sora AI looking like the best investment ever, followed by AI being described as a giant bubble that was going to bring down the economy, followed by that bubble being described as actually not being so bad. This was also the year where Nvidia CEO Jensen Wong took the stage in a conference wearing a jacket that, well, I'll be honest, looks like it came from the prop department from a Mad Max movie. Jesse, let's put this on the screen here. I mean, dude, you're a computer scientist. You were in a racer jacket. I love it. I'm here for it. What I'm trying to say is a lot happened in the world of AI in the year that just ended. And the key question that I've been grappling with is did this year end up being a great year for AI or a terrible one? I would believe either answer. And so much happened it could be really hard to try to keep it all straight. So here's what we're going to do today. We're going to try to get an answer to that query. To help me in these efforts, I've invited to join me Ed Zitron. I think one of the big missed stories of AI in 2025 is Zitron himself, who hosts the Better Offline podcast and writes the where's your Ed's at Substack. He rose to become, I think, one of the more informed and important AI commentators out there. The secret to Ed's success is pretty simple. He just does his homework. Like he actually talks to sources. He talks to reporters, he reads earning reports, he gets leaked information, he talks to people within these companies. He puts together the pieces. Old fashioned shoe leather reporting on what's actually happening with these businesses as opposed to reporting on the stories these businesses are telling about what their technology may or may not do. My honest opinion, or at least my humble opinion, I think he's probably the most important AI commentator that you haven't yet heard about. So Ed is going to join me and what we've done is we've pulled the biggest AI stories of 2025, one per month for the entire year. We're going to go through them in order and Ed is going to help us make sense of what was going on behind the scenes and what these stories actually mean for the AI industry writ large. We'll end up with a conclusion of just how good or bad this year actually was for AI technology. So by the time that we are done with this episode, you will be more or less fully up to speed with where we are at this moment in the world of AI and what is likely to happen in the near future. All right, so let's get in this episode. As always, I'm Cal Newport, and this is Deep Questions. Today's episode was 2025. A great year or a terrible year for AI? And we'll get right into this after the music. All right, so, Ed, we got a lot to figure out. I gotta point out something first, though. Okay. This is something I don't normally do, but for those who are watching, I put on a jacket.
B
Nice.
A
To try to compensate for your English accent.
B
A jacket.
A
I think it's gonna make me look a little bit more scholarly and erudite. That's my. That was my strategy.
B
I'm. I'm wearing a sweater that I've worn once, and I'm like. I guess I'm warm, but I look weird, but it's fine. That's. That's my bit.
A
Yeah, but you sound. You know. But it sounds.
B
I sound British, and I can't hide that.
A
Yes, and so that. That gives you an advantage on me, but I think my. My blazer will kind of balance it out.
B
But, yeah, I think we should be good.
A
You're wearing a. You're wearing a sweater in Las Vegas, though. So that is. That should take points away.
B
It gets cold here. It gets cold here sometimes.
A
I don't believe I went once in July. I'll never believe anything else.
B
Yeah, okay, that. I can understand that.
A
All right, so we're gonna try to figure out what the hell happened in 2025. Right. It's you and I both recovering. AI. In that year, it felt like all the things happened. There was no quiet period in that year from the AI front. And so what I wanted to do is go through month by month and hit some of the big headlines. And you and I will try to figure out what was that. Was that good news or bad news for AI? What actually happened. So it's going to be like a trip down a sort of frustrating memory lane. All right, let's start in January. I actually forgot that this was in 2025. I thought it was earlier, man. It was a long year. All right, in January, we get Deepseek here. Is Deep Seek, the Chinese AI app that has the world talking. Let me read the first sentence of a BBC article from that period. Deepseek, a Chinese artificial intelligence startup, made headlines worldwide after it topped ad download charts and caused US tech stocks to sink. In January, it released its latest model, DeepSeek R1, which it said rival technology developed by ChatGPT maker OpenAI's and its capabilities, while costing far less to create. This was like a huge deal that no one talks about anymore. Explain to my listeners what the hell is Deep Seek?
B
So Deep Seek was a really interesting one. I remember I was on a plane, I was, I was just. It was. I just got started to move back to New York and such like, I spent a lot of time there and I remember reading it about this thing and what it was was that it was a model that was trained for less money than other American models. So American models that cost like fifty hundred million dollars or more to train. Deep seek apparently cost $5.3 million, I think to train. It's really weird because it spooked the entire market. Like everyone freaked out. I remember thinking, yeah, I remember thinking this is an obtuse story to freak people out. Like it was just like even trying to explain because I did, like a lot of media at the time, I was explaining it to people. I was shocked that people even had any interest in model training. But the big thing that spooked people was it was kind of the thing that shown a spotlight on the Nvidia problem, which is that Nvidia is like the only company really making money in this era or just. And I think the people started to realize, oh crap, our entire stock market is based on that. And it also made it clear that all the American model companies don't really give a crap about any kind of efficiency or anything. And the reaction to it was great. Sam Altman suggested we ban him. That was my favorite. But they were like, oh yeah, the sneaky Chinese are going to. It's because they might be able to see inside things. We can't possibly trust them. What was really good as well was part of that. I literally was just reading about this yesterday. Part of what was funny about it was part of OpenAI's complaint was, yeah, they might do IP theft. It's like, no, we only let American large language models do that. We couldn't possibly have the Chinese take away our plagiarism machines.
A
No, we are the world leaders in plagiarism.
B
Yeah, exactly. We can't have the Chinese steal our things. That's our job. But what was also interesting was they were like, should we sue them? Because there's a process called distillation where you basically take another model's outputs and you use them to train another model. That's a very truncated version. And it was. They used ChatGPT outputs to train deep SEQ and that made people pissy. The other thing was, was it was a reasoning model, the R1 model. And OpenAI had at the time only been out for a few months with its Reasoning Model 01.
A
Yeah, that was a December 2024 release if I remember.
B
I think it was September or September. Okay, it was September because it was the run up to that was this whole thing. People like, oh, it's called Q Star, it's called Strawberry, it's going to change everything. It didn't change anything. It really, it actually, it did change something. Reasoning models gave them more excuses to burn AI compute. But yeah, this whole thing was great for me. I did a bunch of media hits about it, but it was peculiar because it was like quite a nuanced story. And then you saw all of this xenophobic stuff being like, oh, oh, well, the Chinese are, they're lying. They're lying. They put out a paper about this. They showed people how it was done because they trained. They had to find a cheaper way to train because they only had. They had, I think, I forget, maybe they had 800 chips, they had quite old GPUs. And the thing is it wouldn't talk about Tiananmen Square. And people were like, oh look, this is proof that it's bad. It's like, yeah, it is bad. It does that. But are you shocked that something that came out of China had censorship?
A
Yeah, but what's, but it was. What's amazing about the story though is it went away. Like, I mean all the points you're talking about are fair points. And like the biggest destabilizing point for the industry was this idea of you don't need the very largest data centers, you don't need like the custom AC Microsoft 40,000 GPU data centers to build really useful language model based AI. But those big companies are dependent on the idea that only they can do it. And so it was almost like people didn't want that to be true. So we just forgot about it.
B
Yeah, we memory hold this bad because at the time it was like, sure, because I got asked quite a lot like OpenAI, surely they're going to make a cheaper model now they did not anthropic. Right. Because they, if they, if they said that that was positive, their idea was we're now going to spend 5 million versus 100 million on this, they would then have very little justification for raising so much money.
A
But what about like the nano models? But then it didn't OpenAI do. The thing is they have some cheaper to use models.
B
Cheaper to use. And that's the thing, people love to use this as proof that the cost of inference is coming down. So inference being how an output is done. And they're like, well, the models are cheaper. It's like, yeah, if you sell something cheaper, it's now cheaper for someone to buy. There's no proof that that's actually cheaper to run. And indeed it would have been so easy for them to just say, actually this is a cheaper model, it costs this much. The fact they didn't means that it's still unprofitable. Which is crazy. But the thing is, even Deep seeks models aren't like no one proved that they're profitable to run. And so, but I think everyone memory hold it because they, I don't know, I think the media just was willing to. This was just a narrative that they could just get rid of.
A
The Chinese angle must have made a difference because to me, an even bigger story, which I didn't really know until talking to a source at the end of 2025, the biggest story that's not being talked about is if you look at coding agents and you look at Cursor in particular, to me, the big economic story was the fact that Cursor at some point quietly just said, we're going to train our own model. We don't need a frontier, we don't need a frontier model. We'll start with open source weights and train them themselves. Now, whether or not that model is profitable for them, I think that opens the door.
B
Cursor has been working on their own models forever. Yeah, ever and ever. Maybe the reason they, they raised $2.3 billion. So maybe the plan is for them to train their own one. But it's. They also, there was a story back in September that Tom d' Artan over at Newcomer, apparently someone said that Cursor is sending 100% of their revenue to Anthropic. So they're still, they're one of Anthropic's largest customers. So it's, it's really interesting though that they're trying and they've gone very hard on composer and things like that. So maybe they are trying that for real now. Maybe, I mean they're capitalized. But the question is to what end is it more profitable? Because if it ends up just being unprofitable and they don't pay anthropic, that would also be very funny.
A
No, I mean I think the future has to be and we can get into more later if we need to. I think the future is really going to be small models, models that fit on machine, right?
B
Yeah.
A
The only thing that's profitable is I'm not spending any dollars to run inference because this 2 billion parameter model which is only really trained to do the very narrow thing I needed to do which is like understand your spreadsheet program and help you do it can run on your phone, it can run on your chip, you're paying for electricity. Like that's got to be the only way that this is profitable. But those companies cannot be large companies. Now you have 10,000 smaller companies instead of OpenAI as the new Microsoft.
B
Yeah, I also think the small language model stuff because small language model is just the large language model with less parameters. And while it is possible to do edge, I just wonder how useful those edge like you can run on device but how long does it take to run a device? Nvidia put out something called DGX Spark box thing that can run large language models. The question is at the end of all this, is it worth it? Like is spending maybe it is worth three grand, 12 grand, whatever for one of these machines. But the companies building these models are not optimizing for that. They're not building models with that in mind. Nvidia has built that box so they have something else to sell. It's been in the works for a while but it's like we don't. No one. I can't find anyone who has run client side who has used it like in that manner, who's like oh, I do all my coding but I do an on device one. I'm sure they exist but the fact there isn't a growing community of that suggests that that might not be viable either. But I think any future large language models will have to be on device. It's just the question is does that happen at any kind of scale?
A
Yeah. All right. So other thing in January was agents. I tracked this down recently. This is when the chief product officer of OpenAI said the quote that then got translated By Axios into 2025 is the year of AI agents. Axios does this by the way. I don't know if you've seen this as a reporter. It's really a pain they invent a quote, they paraphrase what someone said into a better form, and then we'll say like, this headline is 2025 is the year of agents. OpenAI CPO says. You would assume that means that the OpenAI CPO said 2025 is the year of AI agents. He did not. Now he talked about 20. He said things. That was less quotable. They did the same thing with the bloodbath. And Dario Amade, he never actually said it was going to be a bloodbath, but they had a headline that said, this year is going to be a bloodbath. Dariel Amade says. So anyways, I watched the.
B
No, Axios had a quote as well where it was like, this is proof that AI is taking jobs. Then you read the study and it's one line saying, yeah, we kind of see some effect. It's. It's very frustrating because it's. It helps. It's just marketing. It's not helping. Sorry, I'm just.
A
But this is what. Yeah, I know. It's. It does. It does. I actually told. I was. I was talking. I was talking to. Speaking of AI skeptics, I was talking to the Gary Marcus not long ago, and he happened to be on his way to do. Do something at Axios. And I was like, you got to tell him to stop doing the headlines because I keep getting dinged by fact checkers afterwards. Like, we cannot find evidence of this. All right. But anyways, early in 2025, this is when we got agent excitement. And I think this kicked it off. And then around the same time is when Sam Altman wrote a blog post that said they're going to join. Yeah, Reflections, which was probably going to. Agents will probably join the workforce this year and material impact their output. So why. Why would. What are. Why did they start talking about AI agents sort of out of nowhere in early 2025?
B
Well, because they needed something to keep selling this crap. And they launched Operator within January, I think maybe that was February, and it didn't work. And this is a failure across the board with the. The media. They all went, yeah, Operator. It can take actions in a bra. No, it can't. Wait. Okay. It can take actions in the same way that if I just throw a brick, it will take. That is me, I don't know, playing a game, if you consider it. Like, you can take abstractions from abstractions all you want. But, yeah, they. It's this. It's just marketing. It was marketing and mythology that agents. Because agents, much like the term AI, is A marketing term. And that's an Emily Bender, Alex Hannah quote. There is this thing of agents conjure up this image in your head of like, oh, an agent that goes out and does something for you. Now, agents going back to 2023 literally just spent chat. Bottom like, that was what it originally meant. But agents within this era were meant to. Meant to be digital labor. And I take that from Marc Benioff and Salesforce and Agent Force, where they're like, oh, it's digital labor. Sam Altman's comment was. And he always says may or probably, but it's like agents may join the workforce. Egregious lie. Yeah, egregious lie. There was no proof at that time that it was even possible to do it. And guess what? Where we are today, it's not possible either.
A
We do have a coda, because at the end of 2025, we'll get to a news story where, spoiler alert, OpenAI takes all their resources away from agents because it wasn't working, but they were excited about it. I found the Benioff. Some Benioff quotes, by the way, speaking of egregious. He didn't say probably joined the workforce like Sam did. He said it is going to. Not only are they going to revolutionize the workforce, they're going to create two to five trillion dollars in economic activity. Two to five trillion dollars from agents. Wow. Yeah. So, you know, almost there. All right, so agents become a big thing. What were they? To me, I always think, like, what is the. There's a different flavor that AI companies are typically pushing. Like, they have to have a flavor of excitement because of the investment train. Where were we? I don't really remember, like, coming out of 2024, that was more what, like a AGI superintelligence. Like, there was. There was. They were pushing a different message back then because I remember thinking the shift towards agents and therefore a shift towards we will just be in the workplace helping your bottom line. That felt like at the time, a shift towards a more pragmatic vision.
B
I don't know if I agree. I think at that time, that was when you started hearing people say coding agents.
A
Yeah.
B
And coding agents was their favorite one. I'm sure you have some stories coming up where coding agents. Oh, you can get them. And later in the year, anthropic spreads and bullshit around this as well. Where it was. They go out and do things autonomously for you. That was the whole thing that was being pushed. I know because I read every agent story because I found it so repulsive. The idea was that they clearly in the last year, in 2024 they kind of squeezed all they could have out of chatbots. Like no one was like finding they couldn't do more things. They weren't really sure what to do. So they went, agents are coming.
A
Yeah.
B
And what will they do? What do you want them to do? Because they might might do that they can't. But they might. What if they did? Wouldn't that be good? Please pay me.
A
Well wait, but. So I think by the time this interview airs I have a piece out on agents I talked to. I couldn't talk to anyone in the industry but I talked to someone industry adjacent. Someone who made the main benchmark you use to evaluate coding agents and like so here's the story.
B
SWE bench.
A
Terminal bench. Yeah. So I was, I was going deep on what are these agents? And here's what I learned. Okay. So there's, it's, there's two ways AI is helping coders and they get mixed up. So there's sort of the tab complete way which is. Goes back early codex 2021, even pre chatgpt which is it's all based on one shot queries to a language model. So that is like I'm, I'm writing some code right now and I want it to finish. I'm trying to write a function to do something. Just finish this thing immediately that I'm writing and that is powered by you make one query to a language model. So behind the scenes it's here's the code that we've written so far. How do you think I should finish what I'm writing? So that's right in the sweet spot of LLMs where you're trying to complete.
B
Or autocomplete because that's basically how they work.
A
That's basically how they work.
B
They are predicting the next token.
A
And programming languages are highly structured so they're very predictable. So that's. But that's been around pre chatgpt but those work pretty well. Those are now integrated into most development environments. Is my students. I'll call it tab complete. You just oh I don't want to.
B
That's cursor calls it that as well.
A
Yeah.
B
I don't know if you'd say they work really well. Carl Brown from the Internet of Bugs describes it as makes the easy things easier and the hard things harder.
A
Fair enough. It makes, it makes intro computer science problem sets. Yeah.
B
Which still has utility which like that's.
A
Something when I talk to real programmers the main thing they say is useful is they don't have to look up interfaces for function and libraries. So you're like, okay, I can just tab complete a function call here. Oh, this is all the parameters. Okay. Otherwise I would have had the Google stack overload and whatever.
B
But the thing is with that though is that's useful. But if it's querying libraries, could it not get that wrong? I mean, you still have to check its work. Maybe that is quicker.
A
Yeah, it is. Yeah. Yeah. So caveat, I'm tour. Then you had agents emerge, right, which can do vibe coding, right? So like agents was more like, I want you to create a prototype of a dashboard. I want you to add this to a personal website and it does multiple steps. So it's making multiple queries to an LLM. So it'll ask the LLM for like, it'll explain what's going on, here's the tools available, what's the plan? And then it'll go step by step. Okay, here's the output of that last step. What do you want me to do next? So it's a program that's executing. That's what makes it an agent is that it's executing multiple steps, each of which is based on its own LLM query. What I learned from the people in the field is in one sense this worked really well in that it can vibe code. If you say, I want you to produce a prototype of whatever, it could actually get through multiple steps and produce a prototype of whatever. It turns out it was pretty good at this because all of the stuff you have to do to create a computer program you can do is text based commands in a terminal. So the tools that the LLM has to work with were text based commands. All of these coding agents operate in a text based terminal environment. So on the one hand like it could do this vibe coding stuff pretty well in the sense of like it could actually produce a program that more or less did. What's that? I gotta push back, Cal. But wait, I'll let you push back in a second cause I'm going to push back on myself first. But on the other hand, these things that were being vibe coded have no economic utility. So like in the abstract, right, if you're like, can we have a multi step, multi step execution, call an LLM bunch of times to do a bunch of stuff on its behalf and on the other end end up with an updated website or a web based dashboard that like more or less does what you say, that it can do that. But the economic utility of that is very limited. All right, can it do that?
B
Can it do that? Because that's the thing. Vibe coding I believe is one of the greatest frauds of all time. Because if you go on like retplit or lovable or what have you subreddits, you can see people struggling with basic things. While it may get one thing right once, it's not going to get it done every time. And on top of that, Vibe coding does like if Vibe coding is, I am a non technical person building software, it is a lie, it's a fraudulent thing. Because you cannot just vibe code without knowing stuff. You have to know how to read code because something will break. You will not make stable or reliable or secure software. Or on top of that, the multi.
A
But I have seen it, it does create. If you need. There's like a real use case, right. For someone I know if it really is just like I want a web based interface for the silent auction for my kids school and I don't know how to program like it can do that.
B
It can do, can maybe do that or a language model component. It's like will this work this time?
A
Oh yeah, it might take you, but you can work with it and finally get that thing to work without really having to know. But you do kind of still have to learn a lot of this Stu. Fair enough. Because the people I've talked to doing this without coding background, they now know a lot about setting up accounts on these different like AWS clouds and different coding environments. And you do have to learn. Yeah.
B
Which kind of at that point that's not an agent at that point that's just, that's just generative code that may or may not work. And then you need to do all the infrastructural things. It gets back to the thing of what is a software engineer. And it's like a software engineer is not just writing code. I just think that even now, and I'm not saying this is a criticism necessarily even then the marketing is so powerful that even we fell into that trope of like, well, it can do this. Can it? Can it though? Can it do it every time? How much, how replicable is this process? How realistic is this? I think you're right though. It's like it has some use. And I've heard people say this as well with like MVPs, like you need to fudge something together for an investor. You need to do it quick and dirty. It doesn't have to be perfect, but as to kind of resemble the form and function I've heard people do that, and it's great.
A
That's. Yeah, but that. That's what I've heard. And I've heard dashboards. People like dashboards. So we want internally. So, yeah, okay, we're in January, man. Okay, we got to roll. All right, we're going to take a quick break here to hear from some of the sponsors that make this show possible. I want to talk about our friends at Express vpn. So I just heard something mind blowing. Netflix has more than 18,000 titles globally, but only 7,000 of those titles are available in the U.S. you're missing out on literally thousands of great shows unless you're using ExpressVPN. You see, in addition to the world class protection they give for your Internet activities, ExpressVPN lets you change the location from which the Internet thinks you are coming from, which means you can change where Netflix thinks you're actually located in the world. ExpressVPN has servers in over 105 countries, or in exactly 105 countries, rather, and all 50 United States. So you can gain access to thousands of new shows and never run out of stuff to watch. So, for example, when I was in London recently and I was using Netflix in the hotel, there's shows there that we don't get here in the US Like Top Boy or Poldark. If I wanted to watch those shows right now, it would be super easy. I would just have to open the ExpressVPN app. Select United Kingdom Refresh Netflix. Boom. Now I'm seeing the United Kingdom shows. Now. If you're going to use a VPN, use ExpressVPN. It's the one I recommend because it's easy, it works on all devices, and it's rated number one by top tech reviewers like CNET and the Verge. So be smart and stop paying full price for streaming services and only getting access to a fraction of their content. Get your money's worth@expressvpn.com deep. Don't forget to use my link@expressvpn.com deep to get up to four extra months of ExpressVPN. This episode is also sponsored by BetterHelp. This new year, you don't need a new you. You just need to feel lighter. Think about it. Every January, we're told to add more. More goals, more hustle, more change. But what if feeling better in 2026 isn't about adding? What if it's about letting go? Letting go of what's been heavy or stressful or holding you back. Therapy, with better help can Help you see what's been weighing you down. With a licensed therapist, you'll gain clarity, perspective, and emotional space for the possibilities ahead. You don't have to reinvent yourself to move forward. You just have to make room for the lighter, truer version of who you already are. And if you're considering therapy, consider BetterHelp. With over 30,000 therapists, BetterHelp is one of the world's largest online therapy platforms, having served over 5 million people globally. And it works with an average rating of 4.9 out of 5 stars for a live session based on 1.7 million client reviews. Make this year the year of you letting go of what's heavy. With better help, you can't step into a lighter version of yourself without leaving behind what's been weighing you down. And therapy can help you clear space. Sign up and get 10% off@betterhelp.com deep questions. That's better. H E L P.com deep questions all right, let's get back to my conversation with Ed. All right. February. Oh, I know. So it's not going well for AI so far.
B
I mean, at least not great.
A
Well, let's get to February. February, I think, was the quietest month of 2025. There were two models released that I don't remember at all. And I want to see if you remember these at all, too. The two models that were released in February 2025, Gemini 2.0, and this one I really forgot. OpenAI's GPT4.5, which I ended up learning about later. This was them. Tell me if I have this right. This was going back to the scaling, you know, the scaling article I wrote. I talked to you for that. This was the result of the project they started right after GPT4, where they said, we're going to make the model 10 times larger. We're going to make our data center 10 larger, and the result is going to be HAL 9000. And it was it. And it was the like, oh, crap moment where they're like, oh, it was. And this is what they eventually released out of that, I think was four or five.
B
Yes. And I absolutely knew that this was coming. So I have the tweet up and I want to read just a little bit of this because it really tells a beautiful story.
A
But this is your. What are you reading? Your tweet from February, Sam almonds From. From February 27th. All right.
B
Okay. GPT4.5 is ready. Good news. It's the first model that feels like talking to a thoughtful person. To me, I've had several moments where I've sat back in my chair and been astonished at getting actually good advice from an AI. Bad news. It's a giant, expensive model. We really wanted to launch it to plus and Pro at the same time. But we've been growing a lot and are out of GPUs. We will be adding tens of thousands of GPUs next week and roll it out to the plus tier then. Hundreds of thousands of GPUs coming soon. I'm pretty sure y' all will use every one we can rack up. This isn't how we want to operate. But it's hard to perfect perfectly. Perfectly even predict growth surges that lead to GPU shortages. A heads up. This isn't a reasoning model and won't crush benchmarks. It's a different kind of intelligence and there's a magic to it I haven't felt before. Really excited for people to try it.
A
Yeah.
B
This was when you're right. This is when people started going, hmm, hmm. I don't know about Clammy Sam Altman.
A
They're starting to get a little bit worried. Yeah.
B
But want to hear magic?
A
It's magic, though. Magic. It's magic.
B
It's so good that I can't tell you why it's good.
A
It's going to be magic. Yeah. So. So cool.
B
People give him billions of dollars.
A
So from what I understand was they. So they based by reporting for an article we'll get to later. They knew about a year before this that they were in trouble going into summer 2024 for sure. This Project Orion, their sort of next large model training after GPT4 was not generating the same leaps in performance that they had seen before. And so this became a problem. This is why my understanding is in the fall of 2024, tell me if I have this more or less right. They began switching to talking about things like o reasoning models. Models that were tuned. So they were on. Not even on this 4.5base. But like I think those original ones were actually tuned off of the GPT4 base. Right. So they were taking.
B
I. I think so. But Orion was such a mess. There was a Wall Street Journal story towards the end of 2024 where it was like, it's costing a bunch of money and it keeps. It isn't getting better. Yeah. And I think that they were just pure scaling.
A
That was their last pure scaling play is they did the exact same thing they had done for GPT4 and they're like, let's just do that bigger. And it was so that's expensive because that's a big model and it just wasn't getting much better. And that's why my understanding is they switch towards these tuning things because now they're like, well, what we'll do is we'll tune an existing model to do well on different benchmarks or give them specific features and talk about those particular features. So reasoning is what really matters. Not this model is just clearly much more better at everything. That was kind of the GPT4 experience for a lot of people.
B
What I think it is is it's test time, compute. It's just reasoning, as in, like, instead of I ask it to write a fanfic about Scooby Doo doing Tiananmen Square. Instead of it just burping that out, it breaks it down into steps of what is Scooby Doo? What is Tiananmen Square? And so on and so forth. Yeah, and then. And that is the only way they start. They were seeing reliable benchmark improvements.
A
Wait, so walk us through that. So you would, you would. You would query the LLM first to be like, kind of break down the user's prompt, break it down into multiple things. And then they would make multiple queries to an LLM on different parts of this and put it all together at the end.
B
It's a little simpler. It's usually with an LLM before reasoning, you would ask it. This is very simple. You would ask it a thing. It would spit out an output. Instead. Here, when it spits out the first outputs, it's not sending them to the user. It's actually taking a query and saying, what is the user asking? Here are the steps. And this is all output tokens. So it's expensive.
A
Yeah.
B
But it says, okay, these are the things that I think the user wants to do. Time to generate something for each bit to make sure that I'm doing it right. And then the output happens. This allowed for improvements on benchmarks and it had good returns on coding in particular. It also chewed up way more compute. And this helped everyone because the benefits of just training models by shoving a bunch of training data into them, those are. We hit the diminishing returns at the end of 2024 as well. On top of that, there's the post training aspect of basically correcting, correcting behavior, saying, this is a good output, this is a bad output. That's also where they saw it. And actually I realized I'm getting ahead of myself at GTC in March. So the next month. Jensen Huang.
A
We're jumping ahead to March now. So tell us about this.
B
This will be our so in GTC in March, two things. I don't know if you've got all the Nvidia stories because there's some weird stuff, but Jensen Huang on the big screen showed like that pre training. So the shoving the data in. We're past that era. We're into post training and inference.
A
I think we. On the episode, on an episode where we used some audio from you as well, I showed, I think, exactly that part of his speech. So that was in March of 2025. That was his big conference, right? Yeah. And that's where he was. Just as an aside, because we talked about this in the intro of the show today. Why does this computer scientist in his like 50s who makes graphic chips with glasses insist on wearing jackets that seem like they came out of the prop department for Mad Max Fury Road? What is going on?
B
Hey, I will defend. I had the menswear guy on my show to talk about Jensen Huang's jackets. They're sick.
A
Yeah, they got zippers everywhere.
B
I mean, yeah, he really needs to stop doing the racer jackets, though. Those don't. Those don't look right.
A
But the GTC jacket was cooler because it looked like a psychedelic alligator skin. But like he often is wearing like race car jacket, like motorcycle racer jackets.
B
Ones with all the zips I can start doing. I love leather jackets. I can't pull off the races, though. But nevertheless, he also, during that put up a big. A big picture that said, we've so we've it heavily hit. He was like, we've done this many hoppers that we've shipped and now we've done 3.6 million Blackwells and ended up in an analyst Q A having to correct himself to say, oh, I didn't say shipped. I meant ordered. And Also, it wasn't 3. It wasn't 3.6 or 3.2 million. It was actually half that because each GPU has two GPUs in it. Because it's. He counts by the die. It was when you started to see Nvidia start doing their kind of riddles.
A
Yeah.
B
Where it's like, oh, we didn't ship. We. We sold. And they've been ordered from four of the last hype, the largest hyperscalers. And it was interesting because it was the last. I think that March one actually brought Nvidia kind of back to life a bit because looking at the stock and the time, they were kind of trundling and trembling and then they fell down towards the end of April. So this was an attempt to restart the hype cycle. But what was, what's interesting as well was all he did was basically say, yeah, we're going to have even bigger, more huge GPUs and everyone will buy them. We've sold so many and we love selling them and they're so good and they're so expensive and they are well.
A
And they were selling a lot. So like from their perspective, they were. But he loves. But the big other thing of that speech, if it's the one I'm thinking about, right, is that that's where he explained for the media this way. I took in the analyst. Why like the Wall Street Journal's coverage from the year before of like on the information's coverage of Orion struggling, this or that wasn't a problem. And he made it very clear, he was like, look, we were in a scaling era and that made us this good. And now we're in the post scaling era where we do tuning that also requires a lot of GPUs and that's going to keep us going. So he just sort of explained it in a way that I don't think it had been so clear before. And then we were sort of off to the. The media was back like, okay, we're good, we're on track for things continuing to get. We don't know what any of these words mean, but the graph kept going up even as you pointed towards the post scaling age. So that was I think like the first real explanation of like something that changed. OpenAI wasn't really talking about it yet.
B
They were like, it was a subtle shift as well. Because one of the great myths of the AI being bubble is that all of those GPUs were for training. Very convenient to say that because if it's for training, well, we need them. That's the only way the models get bigger. I realize this is a few months in the future, but there was an MIT tech review piece in May of last year where it said the 80 to 90% of compute is actually inference. So the truth is all those GPUs aren't building better models. It's just running the bloody things. Yeah. And I think that this GTC with Jensen Huang was an attempt to bridge that age to say actually the returns, the bigger benchmark scores that we love to see, they're going to come from actually renting more GPUs just to run the models. But we can make the models better by using more compute. Please buy GPUs as opposed to saying all of the compute that we're going to use is front loaded to build the models. No, we need all this compute and you need to sell these GPU, buy these GPUs even because these models, to make them smart, need the compute to stay smart.
A
I see. So this was bad news for the AI companies and good news for Nvidia. This is why they were pushing.
B
It was for everyone.
A
But the AI companies don't like this because they're saying this is going to make it more expensive to deploy and run these products. That's going to hurt our profitability. Nvidia was saying to shareholders, essentially this is actually better for us as the chip sellers because you need all of our chips just to use the product. It's not like, oh, you need, you know, once we train it then it's going to be cheap to deploy. And maybe at some point, hey, it's now we're shifting to a world where just running the product requires a ton of chips. And so like we're great with the market if anything is bigger for us.
B
The thing is, if you look around the startups though, they love this. They love saying test time, compute. They love it. They love saying test time, compute. I'm sorry, but they really do. They loved it because it was a way of saying, well, we need. Because think about it from a startup's perspective. Startup, if they say I need a bunch of money for training, that's a one off operation. Or maybe a couple times a year, I need a bunch of money for compute, I'm going to need a bunch of money cursor. And this happens, I realize later in 2025, ended up raising like $3 billion that year. They didn't raise that for trading, they raised that to keep running their bloody operations. OpenAI building out all these data centers, building them out. They were doing that because the inference cost of running these models, the inference scale, to actually provide a service at any kind of scale required all the GPUs. So it was really just kind of a cartel operation type thing where everyone. And propaganda as well that yeah, we actually need all these GPUs just because these services are so powerful when the real word is lossy. These services are just inefficient crap piles. They're, they're slovenly. It's like the, the very, almost the opposite of what Deepseek was about though. Deep Seek, I think, I think Deepseek required less GPUs for inference as well. But it's like in the face of that Deep Seq story. It's almost like the American AI industry came up with a reason why Deep Seq was both wrong and actually nothing to think about. Stop thinking about it. Stop thinking about Deep Seek. And they did. By the end of April, people had forgotten about it.
A
It's the F150 strategy. Asia starts producing like cheap, reliable cars and instead of Detroit saying like, well, we'll also have to now create like cheaper. That's what people want. They're like, no, we're going to convince half the country they have to spend $80,000 on a completely souped up pickup truck that's capable, capable of pulling a raptor, baby. Raptor. We need a raptor. All right, so then we jump to April, you got this shift. Maybe you know where this came from. This seemed like a shift out of nowhere. So now we have like leading up to these, this talk of it's like very business focused. It's going to be agents, it's going to be like the age of test time. Computing is going to be the future, this or that. Then April, we get AI 2027 and suddenly everyone for a while is back to talking about AI doom and super intelligent. So for people who don't know AI, 2027 was like a fan fiction. I don't know, it was a story that had animated graphs and here, let me read their description. We predict that the impact of superhuman AI over the next decade will be enormous, exceeding that of the Industrial revolution. We wrote a scenario that represents our best guess about what it might look like. It's informed by trend extrapolations. That's the whole thing, by the way. War games, expert feedback, experience, OpenAI and previous forecasting successes. This scared the hell out of a lot of people, Ed. I know.
B
I just, I just spent yesterday reading this thing and pulling it apart. So I'm, I'm, I know all about that.
A
But the key is 2027 they had a non trivial percentage of the extinction of humankind. I think that's the, that's the headline thing from this, right? That like in two years that could be the end of humankind due to AI extermination.
B
So I want to be clear of who these people are. Daniel Coccolitaljo, I think his name worked at OpenAI for two years on the governance team. He was previously a PhD philosophy student at UNC Chapel Hill. On top of this, Daniel quit. And he quit in the middle of June 2024, claiming that OpenAI was secretive and didn't care about super intelligence. Now you'd think if you. And they was named as a whistleblower by Kevin Roose at the New York. You know what whistleblowers tend to do? They tend to blow a whistle. Daniel didn't. Daniel didn't actually say anything. Daniel had nothing to share other than he wrote a scenario in 2021 that was sort of accurate about what the future might be. And AI 2027 is him. The flipping star Codex guy who was a psychiatrist who named Nick Land, a guy with a theory of hyper racism, is one of his favorite writers. Like the people that wrote this are not scientists. They don't really know anything about anything. The whole thing is written. It's like f thousands of terribly written words, lots of scary numbers. But when you read it, it hinges on one idea. Just one. That in 2025. OpenAI. Sorry, open brain.
A
Open Brain, that's their fictional company.
B
In the scenario, which could be anyone, Open Brain invades the thing called Agent 1, which can do AI research. That is the entire hinging of the piece. Yes, it can do a. Do they define what that means? No, they never define it. So just. Just to be clear, this thing, this thing that was written to scare people, to grift, to help, was the AI safety research nonprofit that's connected to the effective altruists. Anyhow, the whole thing hinges on this idea that they invented an AI that could research how to build an AI they wanted. Yes. That's the entire game. They wrap it in the trappings of finance and technical sounding things and there is a bit in it where it talks about neuralese functions and then cites a meta paper. When you go and read the meta paper, it does not cite anything of the sort. The thing they quote is unrelated. New release. Whatever it was, it's an effective altruist thing. It's from less wrong. It's. Sorry. This thing really deeply pissed me off because I had people calling me. I had people friends of mine, people I love and respect, who were terrified by this.
A
Me too.
B
And that was the intention. Sorry, I'm freshly pissed about this.
A
But you got. Because I read it at the time as well, and I came to the same conclusion as you. It's like, is anyone picking up this whole thing hinges on. They never addressed the question of how do you build a super intelligent AI? What's it going to look like? What's the architecture? How does this work? They just said we'll build an AI that could build a better AI and that'll build a better AI and a better AI and then they'll figure it out. The AIs will figure it out. But as I keep emphasizing, we do not know how to build a language model that can produce a software for AI that's better than what any human can produce. That's not how language models work.
B
That's novel ideas.
A
They can produce the topic type of code they've been trained on. So unless you could train the AI and tune it with lots of examples of real better AI, it can't build better AI. Right. It can't leap beyond what it's trained on. And no one is even close to this. We talked about this earlier. We're tab completing function calls with AI right now so we don't have to look things up and vibe coding buggy dashboards. It's unclear how did you get from that to brand new models of human intelligence that the collective AI community and decades of work couldn't figure out on their own? I'm not quite sure where that leap happened. So I was frustrated by this one. I was frustrated by it as well because it all came down to the same it's all the F argument. Right? So the way I talk about, I rant about this on my show a lot, Ed, but the way I talk about it is what happened was, is that the existential risk community came out of effective altruism in the 2000 teens, which was a community that looked at, we want to look at existential risks that might be unlikely but have really high impacts if they happen, like asteroid hits, pandemics and superintelligent AI because, you know, we're doing our rationalist thing and like the expected value of spending money now on rare things with cataclysmic outcomes is positive. That's basically was the, this was Nick Bostrom's Existential Risk center at Oxford was we probably won't get hit by an asteroid, but because it would eliminate all of humanity, it's actually a good investment to invest now in networks like lookit Asteroids. And so one of the hypothetical risks they looked at with Superintelligence post ChatGPT, they went through this weird sort of shift in their brain where they went from this was this hypothetical risk we were looking at along with asteroids and pandemics, to what if it was actually happening now? Well, if it actually was happening, we're superheroes. We're the ones who like pointed. And there was this shift that happened in that rationalist effective altruism community with the people working on existential risk where they shifted from hypothetical to we're just going to convince ourselves it is happening because that makes us the most important people on earth.
B
You're being. You're correct, but I think you're even being kinder. I think they were waiting for a moment to grift. I think they were sitting there waiting, being like, when is. What's a thing we can grasp onto so that we can start draining cash so we can get a bunch of attention? Look at what happened with AI 2027.
A
I'm sorry, I think Daniel Believer, I think he's. I think I made him a superhero.
B
Cynical grifter. I think he's a cynical grifter. Interesting look at what happened when he left OpenAI. He made this big song and dance with this petition back with Geoff Hinton, who I have some other feelings about. With all of these smart people. Oh, OpenAI is doing such bad things. What are they? I couldn't possibly say. He made all of this noise and the media ate it up. All this whistleblower, this brave guy who came forward. What did he do? Nothing. He didn't share a goddamn thing. He was so concerned, but couldn't say what about. It kind of sounds like AI 2027. Oh, I'm so concerned this will happen. What will happen? I don't know. The agent China Build Steel agent too. China agent Agent China. They're scary, right?
A
The specific thing he was recursive self improvement. That was the thing he was talking about was he incorrectly was saying, we're almost at the point right now within OpenAI where the AI is going to be doing the programming for us. And then we're going to have this takeoff idea that goes Back to the 1960s, this recursive self improvement.
B
I have another word for incorrect lie. I think he is a grifter. I think this whole thing was a grift. I think it's connected to the Star Codex guy. I think it's connected to the effective altruists. You'll notice that those assholes popped up with FTX too. They pop up with everything they've been looking. I don't know, I just. I'm extremely, extremely cynical. I don't know, I just feel correct on this because these people, if they. Because here's the thing. This is my entire feeling about this. If they actually wanted to talk about scary bad things that are happening, I don't know, talk about the Kenyans who are training these models for what, like $2?
A
Yeah.
B
How about you go and talk about all the theft that's happening? How about you go and talk about the gas turbines that were Being spread everywhere. How about we talk about the environmental issues? How about we talk about the thing happening today? It's the same problem with Geoff Hinton. These people want to talk about, oh, what if the computer does this? Which is fine. We should have those conversations. But when you were talking about AI safety, why don't you talk about now? Because if you talked about now, you'd have to do something. You would actually have to take action, take a position, make enemies. You would actually have to do something that mattered. Instead, they do this cynical grift where it's always, whatever you're scared of is just a. It's a couple years away, and we've got to do something now, which involves me doing a speech, which involves my speaker fee, which involves my nonprofit, which involves me doing a panel and speaking to politicians. It's never about today.
A
So how do we understand Hinton? Right? Because, okay, there's this interesting situation where obviously, like, Jeff Hinton doesn't need money. He made a ton of money when they sold his startup to Google. Also, he clearly knows the technology in computer science circles. He was just at Georgetown. He really was the guy in the wilderness that was pushing, trust me, back propagation on these deep networks really can work. You just need the data that was. He knows the tech. I did a thing on my podcast earlier in the year where I was comparing him, talking about risks with Yakowski, and it's really, Yakowski comes out of effective altruism, not a computer scientist. And he's like, LLMs are coming alive and have their own intentions or whatever. Hinton knows that's not true because he helped invent the technology. And if you parse his comments, he's very careful. If you really parse it, he really is saying. What he's actually saying is we made more progress on this research than we thought we would. So it stands to reason the same thing could happen with some new type of machine that we don't know how to build yet. That could be a threat. Right? So he's actually being careful because he knows LLMs are not coming alive or autonomous or this or that, though he kind of merges us together. But I'm trying to understand his motivation, right? Because he knows LLMs were at their smoothing out. He's very careful when he talks about it, that it's, we may invent, so we may invent the machine in the future. Like, we should be careful. It's been more sensationalized the way he's reported or talked about them. Somehow make it seem like the Stuff we have now is dangerous. But is that just influence? Is it just like people want to hear what I have to say. Like why is he. What's going on with Hint in your opinion? For like Hinton's like big push for we should be really worried about AI.
B
I think he wants attention, I think he wants glory and I don't think he wants to change anything. I think he's quite happy with the current scenario. Again, I say the same thing I say about the AI 2027 people, except with more ire. Geoff Hinton is a gifted scientist, A noble. Was a noble.
A
Nobel Physics.
B
Yeah. I don't know what the titles are and he is a scientist and all that but again, Geoff Hinton is this massive microphone. Do you ever hear him talking about the theft, the environment as the primary thing? No, it's always couple years away. What if this happens? Wouldn't that be scary? What if, what if my grandmother had wheels? Then she'd be a bicycle. It's this thing of he has all of this power and attention and so called knowledge. What's he use it for? Nothing to do with what's happening today whatsoever. He doesn't go on stage and say hey this is. These gas turbines are popping up, they're polluting black neighborhoods. He doesn't talk about the fact that he's the turbines.
A
The turbines are for generating the power demands of the data centers. Yes.
B
So what happens? And very simple thing of because it takes so long to build power, what these companies would be doing, Elon Musk famously and Stargate Abilene for OpenAI, they're doing this as well. They have these giant gas turbines that they put out that could be spun up quicker. The thing is gas turbines are sold out. They've been sold out and they have like years long wait time. So they're using old ones which are less efficient and pump out more horrible gas. Anywho, I a non scientist know that and I talk about it regularly because that is a harm from AI. Why doesn't Geoff Hinton, a so called AI safety guy, a guy who cares about what AI is doing, never talk about what AI is doing. He always talks about what it might do and I consider that a grift too. And whether him being scientific only makes it a more cynical grift. They feeling say deal, but she at least has a startup. Even though I think world models are another grift that people are going to move to next. Nevertheless, with Hinton he's always going out there to go, I'm so Scared of the computer. What if the computer does this?
A
He.
B
To be clear, we should have these discussions. Those are valid discussions. That's all he does. He doesn't give a rat's ass about any of this. I think he's as cynical as the rest of them. I'm sure he believes this stuff, but he doesn't give a damn enough about the human beings that are alive today. He isn't actually trying to change anything. He isn't trying to. He loves signing open letters. He loves doing paid speaker opportunities where he gets up and goes the computer. It's scary, but that's the thing. Why doesn't he talk about large language models all the time? Gary Marcus does more for AI safety than Geoff Hinton does. I don't care if people are mad about that. I think it needs to be said. I have my issues with Gary, but at least Gary goes out there and talks about the actual harms. Geoff Hinton talks about himself. Geoff Hinton spreads approximate fear, approximate danger, but never really talks about today because, yeah, we should discuss what would happen if this happens. Sure. But how about if you're so. And his whole thing was he quit because he was worried about what they might do. Why?
A
Why?
B
Like you're so worried. Why aren't you doing anything about it?
A
If you.
B
Are you an activist? Are you going to tell people to bomb date? Like, what is it that you want people to do? And the answer is Jeff Hinton doesn't want people to do anything.
A
Yeah.
B
He wants to sit there and worry about something that might happen without dealing with anything today because that would require him to actually do something.
A
That is an interesting point more generally about the sort of expert class turn to AI safety. It's where there's no specificity. I mean, when you saw scientists leaving the Manhattan Project worried about what they did, they had a very clear program. Right. The Concerned unit. We should test ban. We need to roll back nuclear weapons. We need to create these treaties to do whatever. You're right. It's an interesting point that you don't see. Here's what we need. I mean, Yakowski does, but he's kind of crazy.
B
They all signed a letter.
A
Well, you think we should bomb data center. So I guess there's someone who does have ideas, but he's a little.
B
I think Yud's a scumbag, but I give him more credit for at least saying something. For saying something. Because they love signing open letters. Oh, the open letters that they sign. Oh, oh, we should stop working on AGI now. Don't worry, we haven't started. How about we talk about things happening today again? We can't possibly. So then we couldn't.
A
All right, so then jumping forward to May, this is when the big headline then was Dario Amade. This is when he gave the quote that went everywhere about AI could eliminate half of all entry level white collar jobs in the next. I think he said up to next five years. This was part of a longer interview where he also did this equation of test to general capability. So he said AI was at a high school level, then it became to a college level. Pretty soon it'll be, now it looks like we're getting at a PhD level. So now you can imagine replacing what you would hire a PhD level trained person to do in a job. From what I can understand, that was just referring to math tests. They're referring to math tests that the model had passed and someone had said the problems on this math test, which they had tuned it to do well on, were problems that you might give a grad student on a math test that got extrapolated to AI can do what a PhD level employee could do. I mean, I guess if the employee's job was to solve math competition problems. That's true.
B
That's kind of it.
A
Yeah.
B
Well, to quote Gun toucher from Blue Sky, CEO of Oreo Cookies, the Oreo cookie is as important as oxygen.
A
Yeah.
B
Like, that's everything that Amaday says is just. Yeah, it's like a. It's like a PhD student. And I genuinely think that there is this thing with people. I've heard Casey Newton say things like this as well, where it's like, they're like, yeah, the proof that this is useful is people are using it to do their homework. And it's like, do you think there's a homework goblin in colleges? Do you think that's how colleges are run? You write the homework, the com. What goblin eats it, the goblin pays into the endowment. Is that, do you think we go to college to do homework? Now this is a larger discussion though, because there is a degree of college that's kind of like that. That is a problem. But we didn't need a solution to do homework. We need to fix college. But Amadei, just like Altman, just like all these people just says shit. He just says stuff.
A
Why does he always. He always says, I'm reading the quote here. Why does he always say things like. Amadei said he was highlighting it to warn both the general public and the government of what's coming this Is his shtick, right? More so than Altman.
B
It's the same grip.
A
He's always saying, look, I'm the bearer of bad news. I'm the one who's willing to tell you straight. But it delivers the same message in the end as Altman's more optimistic messages, which is, this is the most important technology. It's going to change all the things, all the money needs to come to the people who are building it. It gets you to the same place, right? Like whether you're saying, I'm worried about it or excited about it, if you're saying this technology is going to create 20% unemployment, I mean, if I'm an investor, I'm like, oh crap, I gotta be investing in the company that's going to take those 20% of jobs. Right. It's still a very optimistic prediction for Anthropom. Right?
B
It's cynical. It's this cynical crap that people repeat every single time they fall for it. Every time he made a prediction at one point it was like 90% of coders will be replaced in the next six months. I think he did that in March. Didn't end up being true, it turns out. But the thing that I think people need to realize is that these people aren't special. They know. They don't know. Like, they know stuff. But like, oh, it's going to replace 50. It's just like a PhD student. They're just saying stuff. I could make this crap up too.
A
Well, he says here, I think maybe this shows it. He also said there, I'm reading a Fortune article here. He also said there is still time to mitigate the doomsday scenario by increasing public awareness and helping workers better understand how to utilize AI. Okay, so there is a solution.
B
So make people aware of AI and then get people to use it.
A
Buy more Claude subscriptions and you will prevent the 20% unemployment. Yes. All right, we figured. All right. So then, by the way, having just done an article on agents and my cynicism meter is off the chart for that particular interview he was doing because they knew at that point that even just like using the mouse, forget the back end decisions, wasn't working. They were nowhere near this. I guess you could just say, but people are reporting it. All right. We get into the summer. I think the big news in June that got reported was the MIT your brain on ChatGPT study. So that I feel like this was. There was some other research that came out early summer that was poking holes. There's the big Apple paper and then there's the one research at asu. There's a bunch of papers. They were a little too technical, I think, for journalists to pick it up, but they were poking holes in the reasoning narrative, right?
B
They were saying, oh, yeah, the Apple one.
A
Yeah. Like, this is kind of nonsense. It's not reasoning, it's working with. It's just like. Like, there's no generalization of concepts happening here. That's where they took problems that it could solve, and then they made the problem size a little bit bigger, and it catastrophically fell off the cliff. It was like, oh, it had just seen this size of the. It's not generalizing. It's not so. But those were a little bit too academic, right? But then There was this June 10th paper from MIT Media Lab that introduced this idea of cognitive debt. It was like more easy reporters or writers. So I think this made more sense. It was this. They said, we studied people writing with the AI and it made them. It was worse writing, and they learned less and they were dumber. That went everywhere. So what did you. What do you think about that paper? What do you remember about that?
B
I mean, just that it was a moment when it didn't change a ton, but it made people think about this a little bit more. It's the first time I really remember having conversations where people were like, oh, maybe this is actually having negative effects. Because up until then, people still had this weird thing of, oh, yeah, well, this is a way of learning. This will be your teacher. This will be. You could be a student and learn anything.
A
Saul Khan was saying this, right? Like he had a startup again.
B
The. The CEO of Oreo situation. It's like, of course he's gonna say that. Sal Khan has been on the AI thing for a while. I mean, Khan, Mingo, I think they had a Wall street journal story in 2024 that it just got math wrong. Might be what Washington Post, actually. But this. That study, the MIT one, made people. Oh, this isn't teaching people stuff. It made people realize this is actually not an assistant. This isn't intelligence at all. This is a dumbass machine. This is. It can fill in gaps of stuff you already know, but if you don't know it, it fills in the gaps wrong or not at all. And makes you reliant on a machine that doesn't know stuff.
A
I always think it's like the story about Washington D.C. like, what stories pick up. It's not the stories that teach you something you didn't know. It's the stories that confirm a preexisting bias, right? And so I think everyone was like, this has got to be a problem that you're using this the right. This can't be good. And so when a study came out, I don't know that it's that great of a study, by the way. I mean, I don't want to cast a spur. I looked at it a little bit at the time. I remember thinking like, this was not meant to be like a super carefully researched, like, et cetera.
B
But anyways, yeah, it was a little more academic than a blog.
A
All right, so then this is where things begin to get. This is like the real turn is into the summer. We get to August, and we get the GPT5. I think this was a big turning point. I have kind of the TikTok here, right? So right as it came out, Altman, Right before it came out, Altman was doing like a really big. This is gonna change the world. Song and dance. He went on. Oh, God, what podcast did he go on? I don't know, all the podcast. One of the comedians in the Rogan Circle. You know who I'm talking to? Theo Vaughn. So he goes on, Theo Vaughn compares himself to Oppenheimer and then chokes up just thinking about how powerful GPT5 is, like, what it's going to do. He's like, what have I wrought? It was like. Basically what I had in my mind at that part of the interview was the scene in Oppenheimer where Oppenheimer's in the gymnasium and they're celebrating the dropping of the bomb. And he keeps like. You know, it's this, like, cinematic Chris Nolan IMAX moment where he's cutting from that the images of the explosion and the bodies in Hiroshima. And it's like the most fraught moment. What have I wrought? The emotional climax of the movie that was Sam Altman on the ovan talking about GPT5. Then almost immediately. So by August 11, he's saying, AGI is not a useful term. Let's not talk about that anymore. Let's not be talking about that.
B
Let's move away from expectations here.
A
And then I had my article, which was one of several that came out on the 12th. So within five days, and I was actually on vacation, but I remember talking to my editors and saying, this has to come out. This is a big deal. GPT5 is a big deal. I think this is opening people's eyes. So that's when I wrote my what if AI doesn't get much better than this article for the New Yorker, which I Think I quote you in. You did. And there was like the New York Times had a big article right around this time as well. And a couple. So did I. And so did. The biggest was Ed's obviously.
B
No, but I did reporting on GPT5.
A
Oh, you had good report. Let me just point people towards your podcast. So that whole month, I mean you had a couple really good episodes where you got super into the weeds. Not super into the weeds, but like it was the deepest reporting I had heard on the technical side of like how GPT5 worked and didn't work and why it was going to be. That's where I learned about to get these benchmark numbers they could brag about, they had to use these completely cost ineffective inferences which wasn't going to work. They lost their caching. I think you were deep on that. That when you use GPT5 to do this reasoning that did better on the benchmarks, you couldn't do cached versions of your stuff that made it.
B
It's really simple because it sucks. So story from 2023, by the way, Sam Altman says GPT5 underway will substantially differ from GPT4. Untrue. Very similar GPT5. The thing it did that was amazing was that it has something called a router model in it.
A
Yeah.
B
Which means that it would choose the best model for the job.
A
Now they were solving a practical problem that this isn't a. Originally it was supposed to be this will be HAL 9000. And then it became change everything. We introduced so many models in 2025 and late 2024 that really the main practical thing is we'll have a model like choose those for you so you don't have to know which one to use. It had kind of gone from this is going to take over the economy to like we made things complicated ourselves. Now we're solving our own problem by having it automatically route itself. But that or as you reported, even that feature caused a lot of problems.
B
Yeah. So the router model was reported incorrectly everywhere as being more efficient and a way to reduce costs. However, due to a source that I have, I found out it actually increases the cost because when you have a large language model, say you load a query, you load chatgpt, like write some code here. When it does that, the system prompt is you are chatgpt. You are an assistant that can do these coding things.
A
It adds text in front of your. For the listener. It adds a lot of text in front of your text before it sends it to the model to give the model Instructions about how to answer, what tone to use, like, what to avoid, what tools.
B
Like, do you use a Python tool or a web search tool. Now, before, before this router model, what it would do was you would choose a model and it would have that system prompt and it would not have to keep reloading it. It wouldn't have to keep entering that in. It would just have it and be able to cache that and then say, okay, this is what I'm working on because of the router model.
A
And not only that, just to be computer science Y, they could actually. Because it's sequential, when you run things through these machines, it's sequential. They could run all of this through. The systems prompt's always the same. They could run it through and get the state of all these embedding. Like a lot of math and computation. They could cache and then not have to run all that through the GPUs. They know what state everything is in after that system prompt. So they could have all of the effect of having processed that system prompt without having to actually do the inference time for it. So they could cache it in a way that like, okay, we can have a huge long system prompt and not have to pay for it every time we do a prompt. Every time a user submits a new prompt, which was very important from a money saving. All right, go on.
B
Except that all that goes out the window every single time it routes to a different model. It has to do. It has to completely start again. Has to wipe the system prompt every time it is less efficient. My source was telling me, I can't exactly explain how, but source was explaining that it was actually creating more overhead and the people on the infrastructure side were saying, what are we doing? Like this seems like there was straight up people saying, not sure this is a great idea. Like I'm not sure. Like this seems to be creating more overhead. We cannot cache the system prompt this is causing. I try. I reported this at the time and I took it to multiple reports and they went, no, it's not. I had multiple reports, it's not true. The reporters show them the thing.
A
You talk to reporters and they said.
B
Multiple reports, it's not true.
A
Not, not to blow smoke at you, but I will say no one else had that technical story. I remember listening to your Better offline podcast. Why aren't there more people with this technical story? I mean, I had just finished my own piece where I went deep on scaling and tuning and trying to explain that or whatever, which I thought was also being super underreported. But you had, like, a really good technical story. No one else had it. I don't know that I got.
B
I got lucky. I got lucky. But I also talk to sources regularly, and I know what I'm looking for.
A
All right, we're gonna take a quick break here to hear from some of the sponsors that makes this show possible. If you're serious about protecting your time for deep, focused work, but still need to manage meetings and responsibilities in your personal life, today's sponsor is for you. We talk a lot on the show about avoiding the chaos of modern work. But the truth is, discipline alone isn't enough. You need tools that protect your focus. That's where Reclaim AI comes in. Reclaim is a smart calendar assistant built for people who treat their most valuable resource, time, as their most valuable resource. It automatically defends your focus time. It schedules meetings, it resolves conflicts and protects time for habits, tasks, and even recharging breaks. Now, here's the best part. It's not rigid. Reclaims AI dynamically adjusts your calendar around your shifting priorities so you can stay flexible and focused. Deep work won't get crowded out anymore. The average Reclaim user gains 7 extra productive hours a week, cuts overtime nearly in half, and experiences less burnout and context switching. Reclaim AI isn't just another scheduling tool. It's an intelligent time partner that helps you prioritize what truly matters every week, try it and reclaim your time for what really matters. You can sign up 100% free with Google or Outlook Calendar at Reclaim AI Cal Plus Reclaim is offering my listeners an exclusive 20% off, 12 months with discount code PAL20. Get your Reclaim recapped 2025 when you sign up, which is a year in review of how you spend your time across deep work meetings and work life balance with 30 plus productivity stats. That's worth signing up. Just that alone, you got to check it out. It's really cool. So go visit Reclaim AI Cal to sign up for free. That's Reclaim AI Cal. I also want to talk about our friends at Caldera. Here's the thing about aging. When you're a guy, when you're young, you don't think about your skin. But then one day, you look, you wake up, you look in the mirror and you realize, oh my God, I look like a grizzled old pirate captain. What happened? You weren't taking care of your skin. Luckily, this is where Caldera lab enters the scene. Because they have a high performance skincare line designed specifically for men. They've got these great products like the good which is an award winning serum. The eye serum, which helps with puffiness under your eyes and the base layer, which is a moisturizer that you just use every day. Look, this stuff works. In a consumer study, 100% of men said their skin looks smoother and healthier. Now, skin care doesn't have to be complicated, but it should be good. Upgrade your routine with Caldera Lab and see the difference for yourself. Go to caldera lab.com,/deep and use code deep at checkout to get 20% off your first order. All right, let's get back to the show.
B
Like that's really, it's like the moment GPT5 went out, I went to source being like, hey, hear anything? Like real simple. I don't even do much source stuff because I have like other stuff I'm working on. But this one was just like I went to someone, an infrastructure provider and I said, hey, have you heard any GPT5 stuff? And they said, let me check. And I went this and it was just, I, I, it shocked me that no one else wanted to cite it. It shocked me that I had multiple people who were just like, this is not true. And I'm like, I can show you stuff that would prove it. They didn't want to see it.
A
Interesting.
B
It's just, it's denialism. It's because when I, I understand. When you've got editors who are pro AI, when you've got other people you work with who or perhaps you yourself want these people to win and you don't want to piss off their PR people, you don't want to lose your access. I get it. But it's like this is, this is reality.
A
They don't have any, no one, no one has access, by the way. I mean they won't give you access to anyone. I'm into legacy media. They give. No, they give. These AI companies give no access.
B
No, yeah, they, and they Play. I've heard OpenAI plays people against each other outlets. Yeah, I've heard that. I've heard them straight up say, apparently, I'm sorry, I've been told that apparently they will straight up say if you piss us off, we'll stop responding to your emails. Yeah, little worms.
A
Yeah. No, I believe it. But this did change a lot of things. So again, to blow a little bit.
B
More underwhelmed everything, everyone was just like they were under.
A
So because it was hard to ignore, the GPT5 wasn't that different and had these other problems. The thing I saw change. So there's articles the hey, scaling this is a problem. That's why my headline was like what if AI doesn't get much better than this? Like that was like a new idea for people. But the thing that that seemed to really open up was a story that really only you had been covering for a year. So for a year plus you had been actually gathering earnings revenue numbers. You've been looking at earnings reports and you had been making the case for about a year up to that point. The numbers don't make sense on these companies. Look at how much they're spending, how expensive the numbers don't make sense. This cost way more to run than they're getting in revenue. When is the musical chairs game going to stop? Post GPT5All the major publications sent good financial reporters to do these type of stories. So we get, for example the New Yorker had a big in the magazine, a big wait, is this a bubble article? The Wall Street Journal had several, including the one in September. Spending on AI is at epic levels. Will it ever pay off? We began to get really good analysis like comparisons to level three and what happened with like Lane the infrastructure. The New York Times started writing these articles. They had covered it the bubble possibility 0 and then they started covering it multiple times. That's my story of September is all of these different bubble articles. I'm assuming that was all basically opened. The floodgates were opened by GPT5 underwhelming. It just sort of changed the way that people categorize like wait, maybe there is, there could be a problem here. Which by the way I have experience with from my social media reporting back in the day. Just as a quick analogy, everyone thought in the media that I was eccentric for my stances about social media is a problem. We shouldn't be using it. This is not a fundamental technology. This is a real problem. And I was shunned and attacked and people were coming after me. And then post Donald Trump election where he was successful on Twitter it planted the seed of like oh, maybe social media isn't just done for good. And it opened the floodgates. And all of these issues with social media was suddenly open game to be covered well beyond even what I was talking about. This felt similar to me. GPT5 underwhelming opened up all of these. The possibility of all these stories including on economic struggling.
B
There was also a big story around August where it was, I think it came out that AI data center capital expenditures made up more of GDP growth than all consumer spending combined. And then in September you had that insanely Funny story where, as we've discussed, $300 billion deal with Oracle between OpenAI and Oracle, where OpenAI will give them $300 billion they don't have and then Oracle will serve them compute from data centers that are not built yet. And I think that was. That happened and sent the stock spiking and then rapid fire. We had this AMD deal where amd, it was a really funky deal as well. It was, let's see, that came out. That was. I'm skipping ahead to October here, but the AMD deal. But basically by October, mid October, I think OpenAI had agreed to like 26 gigawatts of data centers. And there's just a bunch of funding that happened around here as well. But it really felt like the air had been sucked out of the room.
A
People started going, there was scrutiny suddenly on some of these stories in a way there wasn't four months earlier.
B
Yeah. And it's interesting because even with that scrutiny, can we do October as well now?
A
Yeah, we can be in October. Yes.
B
So I bring this up because even with all that scrutiny and the reason I'm typing is I need to bring up the need to bring up these deals. So in, in. Okay. In September, it was this Nvidia, I'm going to do big old air quotes. Nvidia does $100 billion investment in open AI. Now what I really remember at that time was no one having any details about it and indeed the writing within The Nvidia and OpenAI announcement not really being clear about when things will begin or indeed if any agreement was signed. And I went to multiple reporters, I'm like, hey, look, first of all, have you done the maths here? Because if you did the maths, it would cost OpenAI, like I think over a trillion dollars for the compute and their data center deals. And I put that out fairly early. And then just other people wrote the same headline and did not quote me. Thank you. But OpenAI agreed to a 6 gigawatt deal where they built 6 gigawatts of data centers for AMD and in return would get 10% of their stock. Never happened. Broadcom did a deal with OpenAI, 10 gigawatts of data centers, which we will get to in a minute because some funny stuff has happened. And then this Nvidia deal. Now what was funny about this was I went to reporters and saying like, hey, look, nothing's been signed. This is a lot of money. And also a gigawatt data center takes about two and a half years and $50 billion to build. They meant to start these data centers. The first billion dollars, that $10 billion, even that Nvidia was meant to send to OpenAI was meant to be next year. Same deal with AMD, same deal with Broadcom, so meant to be in 2026. And I went to people, I'm like, hey, this isn't possible. Like, this is not. This is quite literally impossible. We cannot do that. Like, you can't build data systems. And everyone's like, yeah, you know, well, they're working out the crap they've been saying for the last year.
A
It's just like they're working out, they're working out. What's the advantage they get by announcing these deals? Can they mark it up as, like future assets? Does it help with stock? What would be the motivation for stock spikes?
B
Okay, and so Oracle added $300 billion. Their remaining performance obligations. Broadcom added like 50 billion.
A
I want to say, because you can mark this up as expected revenue, which when you're then doing the calculations and figuring out, you're like, oh, this makes this a more valuable company because it's the revenue it has or expected has gone up.
B
And because the markets are, I assume, run by toddlers, everyone believed it. Everyone was like, wow, wow. Number go up so big. Number so huge. Well, number didn't stay big for long and things started to fall apart. And it got to this point where people, even people who are quite cynical, started going, one moment, anyone done the math here? And the FT has stepped up. The Financial Times has been pretty on top of this the whole time, but they really stepped up and did some analysis. And it was just. They also did a trillion dollar story without citing me.
A
Well, but the Brits are more suspect of Silicon Valley.
B
But nevertheless, it was this thing of everyone suddenly starting to do very simple math of like, well, OpenAI's projected to make $13 billion this year and they owe $300 billion. How do they, how do they pay that they're going to lose billions? And the information part of a story saying OpenAI was spent. They planned to spend like over $150 billion or something. Didn't really make sense mathematically at all because the $300 billion. But it was very interesting time, actually. That reminds me. So when the Oracle announcement came out, OpenAI had leaked that they would spend, I think $155 billion or something. But it was days before the Oracle announcement was made. So OpenAI leaked their costs. I'm doing air quotes again because I don't trust any leaks out of OpenAI five days before the Oracle deal so that no one would do the back mouth and go, wait a second, this doesn't make sense. I love watching this kind of like disruptive public relations work. I think it's cool as hell. I think it's good that I'm watching because what's gonna end up happening there is no one gets paid, which as in a few months time in this story we'll get to. But I actually have my own story in October as well. I got Anthropic's Amazon Web Services bills.
A
Yeah. So what's going on with that?
B
What, what's that $2.66 billion spent in three months. Sorry, three quarters. $2.66 billion. And that's just on AWS. And from what I know, they also spend about the same amount on Google Cloud. So Anthropic will probably make, I'm going to say, $5 billion this year. They were, I would think by the end of September, they, between Google Cloud and Amazon Web Services had spent more than $5 billion. So they're just annihilating capital, just burning it down. And it actually leads me to an important statement, which is Anthropic has done. The other thing that Dario Amadei has done is he's framed them as this more efficient company, this company that is more efficient, that doesn't burn as much as OpenAI that spends less on training. But when you look at the numbers, it tells a different story. OpenAI in 2025 raised $18.3 billion other than SoftBank's portion, but nevertheless $18.3 billion. People say anthropic, they're spending less money, they're more efficient. Anthropic raised $16.5 billion. Basically the same neighborhood of numbers. But Anthropic has done such a good job just lying to reporters and spreading these rumors that people believe this. I think Anthropic is as big a crap pile as OpenAI. They're just as lossy, they burn just as much money. And yeah, I mean, by this point, by the end of October, I think most even outlets had begun to say like, oh crap, oh crap, were we wrong? Were we wrong for three years? Did we get, did we fall for it again? And they did fall for it again.
A
So then if we jump forward to, well, the other big story in October is Sora the app, which confusingly is powered by Sora 2 the model, because there's also a Sora 1 model that didn't land, I think the way, I guess OpenAI hoped it kind of freaked out a lot of people, like what are we doing here? Who is asking for this? But was that a sign? I took that as a sign of a little bit of a sign of desperation. This is OpenAI looking at TikTok does $33 billion a year in revenue and like we need money, so can't we do TikTok with AI and like help fill backfill? Right? So like in other words, if you were about to automate half of the jobs in the knowledge economy, you don't need a TikTok clone. You don't need to talk about. Around the same time they also Talked about allowing GPT, ChatGPT, more erotica, et cetera. You don't need that. We're about to create the $3 trillion to mark talked about. Who cares about that? But the fact they were putting that out was start sort of taken as like a. Oh. Type of moment, which I don't think is what they were hoping.
B
I don't even think they thought around. I mean people bought the TikTok thing hook, line and sinker. I think they were just desperate. I think they did key jingling. I think like look, look, you generate videos. Please, please, please keep using this.
A
Don't talk about the Oracle deal. Don't talk about the Oracle deal. Look at the, look at the keys.
B
But they did like an Oracle style deal but with their own software. So you had this situation where Sora, now Forbes estimates that it cost them like $15 million a day or something based on nice sources. That number might be smaller. Sorry, that number might be too small. I have compelling evidence that to run 13 instances of Sora 2 required 900 or 840H200 GPUs. That's 13 instances. Instances. That means 13 generations at once. So that's 10. Like this thing is really expensive.
A
What's the cost? If you want to make. So you want to make sure videos, what do you need? What level? I mean they're still taking a lot.
B
I mean you need to use their.
A
API, but you have to use. You have to have their 200amonth level or above. Or is it more?
B
No, anyone can use Sora, the app.
A
I mean for creating the videos though, right? So they're, that's. You have to.
B
You can create them on the app. You're limited. But if you want to use the API, I think it's like a couple dollars per video and per video just means anything it generates, whether it's good.
A
Or not, which is not. That's not Sustainable. The whole point about TikTok's model, that's brilliant is all of the compute involved in taking videos, editing videos, trying a bunch of different experimentation is all done on the user's phone.
B
They pay for it, but also TikTok loses money.
A
Yeah. Oh, interesting.
B
Like, TikTok is an unprofitable business as.
A
Well because it's just the cost of.
B
Hosting, because of marketing.
A
Just marketing and hosting and streaming a bunch of videos, I guess.
B
Yeah, yeah. It's still an expensive. And also they are poised for growth. But putting it aside, you're still completely right. That's how that service runs. Sora was just an attempt to try. And it's like, I don't know, like, when you ever see a couple that's like about to break up and they're like, yeah, we're, we're going on vacation. It's great. God, I love them so much. It's all going so well. I love it here. And Sora, you had all of these. What was funny was Sam Altman, I think true and on had this point where it's like, it looks like Sam Altman put himself in it so that he would make himself famous. And all that ended up happening was people did like Sam Altman stealing from Target.
A
I saw a lot of Samoan crying or.
B
And it was just. It was weird and bad and it sucked.
A
Yeah.
B
It got a bunch of media attention. A lot of people got scared because that was the other intention. It was meant to give people the sense that this would replace videos in general.
A
Yeah.
B
This replaced social media. It didn't. It obviously didn't. And it's obviously too expensive to run.
A
Yeah.
B
And I think that it gave them the top of the App Store ranking very briefly. And then because it's like everything with large language models, other than in really specific use cases, it is just a toy. A really horribly expensive one too.
A
So then if we jump ahead to November, I summarize November as basically, here's what's interesting to me about it. Multiple models from different companies. GBT 5.1, Google Gemini 3, Anthropic Opus 4.5. No one cared or noticed. Which itself I think is significant. Suddenly no one cared. I mean, there were some Gemini. They cared about the fact that it was using different their own chips. And there were some economic stories there, but no one cared that Opus 4. Five was better at coding agents or.
B
That 5:1, because better doesn't mean anything anymore.
A
The other thing I saw in November was there was kind of a defensive backlash to the Bubble stories. So now you start to get, well, wait, wait, we'd gone too far. Maybe it's not a bubble. We begin to get to like, I think it might be okay story. So that was like, I don't know your take on that. That seemed to happen in November.
B
Oh for sure. And we had. But by this point Sam Altman, Mark Zuckerberg and Jeff Bezos had all said it was a bubble. Like all three of them had said it. But also we had a bunch of people doing stories that were quite literally, actually it's. Bubbles can be good. Actually it's a good kind of bubble. None of these had particularly logical points. And so we had these people trying to work out like crap. Did we. You kind of put it like this. It was like, did we over correct? Oh no, I don't want to piss off the powerful people. LLMs are actually great now, but they're actually bad. And then it got to this narrative of, well, remember the dot com boom? Yeah, remember the dot com bubble. And there were, there were companies left over at the end. And it's like, did you read about the dot com bubble? Because like two, like Lucent got acquired. Lucent did probably the best of them all other than like Cisco and Microsoft who kind of survived Amazon. They've done really well, but it took them a while to get out of there. Amazon, Amazon was interesting as well because Amazon didn't like it. It was within the universe of the dot com bubble, but didn't make all the mistakes that they made. Also the thing with the dot com bubble was, and this is how we get to Nvidia in a minute was just the insane deals like Windstar communications getting a $2 billion loan from Lucent Technologies which would. And in the press release said this make $100 million of revenue. It's like they don't teach you that in business school. Also in the middle of this month I got OpenAI's costs. I got. They spent $8.67 billion on inference. Just inference through the end of September. That was a great story because the FT and I worked on it. But the denial around that was really cool.
A
Wait, that's 8. Go to these numbers again. 8, 8, 7 billion in inference. Meaning like just what it costs to train and run their models. Just to run the models and then what against what revenue for that period.
B
So that was the fun thing. So I also got the revenue share from Microsoft. And the way it worked out was because OpenAI had leaked that by the halfway point of 2025 they had made $4.3 billion in revenue based on the revenue share, because they do 20% revenue share with Microsoft. So Microsoft, I could see what they'd been paid.
A
Just multiply by five or whatever.
B
Yeah. And they made 4.3 something billion through the end of September. Now people then said, but Microsoft pays a revenue share to OpenAI too. I actually have those numbers now and it works out to like $4.5 billion through the end of September. I don't know if we ever find out what happens with OpenAI, but I will say this. Those numbers do not match up with anything else reported. And people did intellectual gymnastics to try and say. They said, oh, your numbers are delayed, They're a quarter late, they're three quarters late. They're accrual accounting. I play to win. I know what I'm doing.
A
Wait, just to be clear for the audience though, you're saying through September you're talking 4 point something billion dollars, probably in revenue against already close to $9 billion in inference cost. Correct. Yeah.
B
Not good.
A
You want the first number to be larger. Yeah.
B
You ideally want those to be reversed.
A
Yeah, yeah.
B
And it's like.
A
So they're operating. Yeah. Which is the issue is you're operating at a loss.
B
Massive loss.
A
Yeah.
B
And these costs increase with revenue. That's the actual problem. It's like if costs were going up, but revenue. If costs were going up this fast, but revenue was going this fine. The problem is, and this was what I saw with anthropic as well, because with anthropic spend, I actually compare the revenues versus spend and it's. It just goes like that. It scales with. It's clear that the more money you make, the more you spend. And there's no real reversing that trend.
A
If you zoom in on a user that's paying X per month. The problem, you're probably costing you more than that for that user. And that's why it doesn't.
B
Well, that's because. And that's the unique problem with large language models is you can't do cost control. Yeah. Augment code. I think in the middle of October or November put out a thing saying they had a $250 a month customer spend $15,000.
A
Yeah.
B
In comp costs. Claude code. There's a leaderboard that called Vibranc, because with Claude code you can actually find out how many tokens you're burning and extrapolate the costs. Someone spent $50,000 worth in one month on a $200 a month subscription. That's Large language models, baby. That's just how the cookie crumbs.
A
Yeah, people underestimate like the brilliance of a company like Google Search is. It is really cheap to run. Like they built their whole. I mean I guess it's like the acquired episode from the fall, but I know a lot of Google stuff, they built a very cost efficient infrastructure. That's what they figured out is like we're dealing largely with text and we can cache most of this stuff and we're moving very little bit. And we could use commodity processors that are idle a lot of the time anyways and it's not that expensive to run and we can get a huge amount of revenue per search done versus we can generate $2 in ad revenue on like $0.07 of. That's why that was a money cash cow is they thought a lot about the compute cost and they were like this can be super efficient. And they built an infrastructure from scratch for Google Search to be super efficient. And because of that it became like a cache fire hose. What you're saying is that's impossible for LLMs because the way an LLM works is you have to fire up every one of those stupid weights and run it through a GPU to generate a single token. The whole LLM, every weight is involved. For every token of every response. There is no cost effective way of doing it.
B
But even mixture of expert stuff still run into the same problem because they're imprecise in how they call up the experts. It's because of the probabilistic nature. But on top of that people will also say, well what about Amazon Web Services? It's the favorite comeback that people have. They burnt a lot of money.
A
Nuh.
B
I actually went and looked in the space of nine years, they spent about $70 billion to build AWS. 70 billion. That's less than half of the cost of OpenAI's infrastructure.
A
Well, and also that scaled with revenue. But the thing AWS was a very, it was a very. I mean it came out of what people know or don't know is like where it came out of is they built this infrastructure for their own computer and then it was incrementally. They could be like, oh well we know how to do this now, why don't we offer this to other people? The revenue curve was like the opposite of what you described for OpenAI or anthropic. It was the more people using AWS, the more money they would make. So you could invest money in this. This is not. I mean you could grow the growth here is not nearly as expensive as building out the infrastructure for AI. Right? This was like, this was more. They knew how to run these data centers. These are more standard data centers. It was more. The software was, the main innovation was in the virtualization software, which once you program it, it's free, it's yours, it's your ip. These were known quantities to build out. And I don't know the details, but I would assume you'd go a little bit into the red. You could immediately, in two years, get back in the black. It was a much more controllable space.
B
And there was a path, there was actually a path to it.
A
And then it started making money. And then like, oh, let's 10x this. Because like we see the revenue. If we 10x, it will 10x the revenue. And it kept Amazon in the black for a decade where they were giving away prime memberships way under cost.
B
Also, the business model made sense. People needed to host websites and apps on the Internet. We still don't have one of those for large language models. We don't have a thing we can point to and say, this is what they do that actually makes money. This is the economic viability of it.
A
All right, so December, looking at last month, there's Disney. Disney.
B
There's Nvidia as well.
A
And Nvidia. Okay, so Disney's easy. Michael Burry.
B
Okay, Disney.
A
All right, let's cover Disney first. And you tell me Nvidia. I listened to you just this morning talking about it. But okay, Disney for some reason puts what they put a billion dollars into.
B
That should cover like a month's inference.
A
Cost into OpenAI for Sora. This smells like to me what I don't like about the AI of the last couple of years. The thing that often annoyed me as much as anything else was the executives at unrelated companies who did not understand the technology, who felt like it made them look cool and forward thinking to be like, we gotta do AI or we're just, we need to be doing AI. Go do AI or you're not going to work here. And you'd like the AI to do what? Like, what's this? What specifically are we making money on? No, no, just we're doing AI.
B
Stop asking questions.
A
We do AI right? There's. If we don't do. And you sound like as a CEO at a board meeting, you're like, we got to do AI or we're going to fall behind mind. And they just leave it there. This felt a little bit like Bob Iker saying, we got to show our Shareholders, we do AI, but at a level that's like we can absorb the loss. Is that right? Well, I mean, they're going to use Sora to create. It made no sense to me. People will use Sora to make custom Disney videos with themselves in it. Like, what? I don't understand what's actually going on here other than Iger can be like, we AI good now. It's about as far as I can.
B
So middle of May 2024, Iger actually said that we need to embrace the change driven by tech innovation, referring to AI, and that Hollywood storytellers needed to. I think that what's happened here is that they wanted to invest in OpenAI. Maybe they were going to sue them and OpenAI. OpenAI just kind of scammed them a little bit. Scammed them and said, oh, yeah, yeah. Well, what if we gave you the opportunity to invest? We're not letting anyone in. And so they agreed to that. And they're going to have 200 Disney characters. And the actors unions are pissed off.
A
They just want the stock. They just want the stock position. So now Iger can be like, look, we're hedged against AI disruption because we have like a non trivial stock position in OpenAI.
B
Yeah, I guess. I mean, it's just like.
A
But they're not using, they're not building tools for film production. It's like in this weird sloppy IP space. Right.
B
And the thing is as well, the first time you have like Goofy doing the introduction of Frank from Blue Velvet, the moment that that happens.
A
Yeah.
B
You're gonna see this shut down. Like, you're gonna. People are gonna be on there day one trying to make Goofy have sex with someone, have Donald have sex with someone. Mickey doing 9, 11, whatever they want. Which they already were doing with Pikachu when Sora came out. This is all they are. And the thing is, Disney's crazy. They already had this problem. There was a Fortnite thing in 2025 where they put a generative AI Darth Vader. Within one hour, people had it saying slurs. I remember the Internet is built to generate that kind of.
A
They put a chat bot behind a Darth Vader character so it could talk to you. And they made him into. Yeah. Giving racial slurs within. Yeah.
B
And it was just immediate.
A
It was giving like really unsettling dating advice to the character. I remember covering the story. Yeah.
B
And it's like, think it's funny as well? Because that's obviously what's going to happen. But again, this is one of these deals where it's like, it's going to happen sometime in 2026. Yeah, sometime at some point. It's always at some point with these deals. Sometime, somewhere, some point. Has Disney actually invested? Is that money? There's a licensing agreement. Is it a licensing agreement? I'll be looking at Disney's earnings when they come out. But it's just a very boring and cynical thing. I think Sam Altman is a good con artist and I think he's good at convincing rich guys to give him by scaring them.
A
Well, two other stories that coda things we talked about earlier in the year, which might color our analysis of the full year when we do. So some of the writers of AI 2027 basically came out and said, well, this is not going to happen anymore. We'll do another one. They updated mid November. It got a little bit too far. They're like, actually it's not going to happen. And then also this was in December where the code red was declared OpenAI where they're like, basically we need to make chat GPT better. And they. One of the things they said after the beginning of the year was this, the year of the agents. In December of the year, they said, we're de emphasizing agents. We need to put more energy on making chat pop. So there's kind of this like tragic coda to the end of the year.
B
It was a leak. It was a leak to the information about their costs. And they were like, yeah, we expect $26 billion less revenue from these.
A
It's like it was an internal memo that was leaked. Leaked. So it was the information got it first and then the Journal picked it up, I guess. Yeah, but yeah, in that they listed like, we have to de emphasize agents because we need to make more money on our core product. Because, well, this is a Google Gemini reaction, I guess.
B
But so this was great though. So Gemini 3 comes out and just before that there was a story in the information where OpenAI was like, yeah, we're gonna. Sam Altman did an internal all hands thing where he was like, yeah, we're gonna have some economic headwinds.
A
Yeah.
B
Around that time, Alex Heath from Sources reported the cfo Sarah Fryer, who we also missed that she kind of hinted at a government backstop, but that was kind of. That kind of went away. Nevertheless, she said that there was slowing growth due to safety features. Then Google Gemini 3 comes out. Google stock spikes. Gun to my head. I could not tell you what's different with Gemini 3. I've talked to multiple people. They're like, it's better on benchmarks. I'm like, okay, but does it do anything?
A
It didn't need the video. That was the issue. Google had been working on their own chips and they trained it on their own chips. But that's the thing. Is that the case?
B
Google's got a lot of Nvidia GPUs. That's a convenient story for them that they leaked. TPUs have not been proved. There was a whole argument between analysts about this. Nevertheless, Gemini 3 comes out and because the media just cannot come up with unique ideas like this. This is big, this is, this is different. Stock go up, number go up. And there was this code Red that you mentioned that gets called. And what's great about the code Red story for the information is it's like. And OpenAI had a plan. Step one, we're gonna make the. We're gonna make chat GPT's responses better. Step two, we're gonna give people to use reasons to use Chat GPT more than other models and prefer it over other models. And three, we're gonna improve the functionality of Chat GPT. To which I ask, what the hell have you been doing all year? Yeah, what have you been doing? What if is it? I think OpenAI is like an adult summer camp. I think that they're all just dicking around, doing random projects. No real management. They're just like, I, I think all of. I think anthropic's the same way. It's like, I don't know what we get. I'm working on a model thing. Sure. I'm also, I've heard multiple stories that you have teams in OpenAI working on the same thing that do not talk. They're just like bumping their heads together. It's like the minions in there. But this code Red happens. And at that point, really, you saw the media shift of, oh, God, OpenAI is bad. I think just everyone was like, ah, wait, does this company lose billions of dollars? Did anyone say anything?
A
Why didn't anyone tell us this?
B
Oh, my God. When those articles come out, I'm going around with a, with a mallet. I'm going to be like Mario and Donkey Kong. It's going to be messy. But that's the thing.
A
Yeah.
B
Everyone was kind of like, hey, OpenAI loses so much money and they don't appear to make enough to pay their bills. Is that good? It's so. And every day there's a news story where I will post it and say, is that good? Because it really is just like nothing none of this ever made sense if you looked at it. But it's like really that you can see the milk is curdling in real time. You can see it happen. And you've had terrible earnings. You've had this Broadcom earnings, Broadcom being the one that was meant to build chips for OpenAI. Now the revenue for that is not coming in 2026. It's crazy. It's completely nuts. Oracle, they, they, I think they missed on several parts of their earnings and a lot. 300 billion out of their 455 billion dollar remaining performance obligations is open AI. And people are like, hey man, how are you paying? How are you getting paid for that? Where's the money coming from? Because like you need money to, to, you need money in your business. That's how you make money. And no one has a good answer. And now Oracle has delayed those data centers. So it's like, right, because they can't.
A
They can't afford to build them.
B
They can't afford to build them. They raised $18 billion in bonds and they're trying to raise another $38 billion with vantage data center partners. It isn't clear if that's going to happen. The credit default swaps, so betting against Oracle saying they might default are at the highest they've been since 2009 is, it's. The era of smiles is beginning. It's really, it's, it's dark out there for them. But I'm laughing, I'm having a good time.
A
Well before, so before I get your final take on the year, let me just get your, your opinion on your official answer on this because the, the number one thing I hear from people who don't think my coverage is too skeptical of AI, like the people who are really AI boosters, the number one thing they say is these details don't matter. Cal, you were wrong. You're really underestimating the likelihood that there's going to be these quantum leaps. They're going to come alive. They're going to be. It's AGI. So it's going to be so transformational. Why are you talking about 4.7 billion versus 8.7 billion? It's the future of humankind is about to change. And whenever I do an episode on consciousness or super intelligence and why as a computer scientist, I say this is bunk. I mean, my toaster mice will come alive. It's like, no, no, no, you're wrong. And they really get in the weeds of trying to argue with me about. There's this other Story of these models are on the precipice of transformational change in the very definition of intelligence and AI and what machines can do. Have you picked up, you cover this as closely as anyone. Is there any inkling from people who are in these companies, the analysts who are analyzing these companies financially? The, the, is there any inkling or any care, any attention put to this idea actually put to it that no, no, this technology is going to make a leap into being like intelligent or conscious and it's going to solve all the problems. I know there was some. That was the way they used to talk about it. But just to clear the decks, is there any conversation about that actually seriously happening anywhere tied to these companies? No, I just need you to say that I needed that clip to be able to give the people there.
B
It's just no. And my evidence is. All of the stuff we've been talking about, their evidence that these are getting exponentially better is fairy tales. It is. Well, what if this happens? If a frog had wings, it could fly. It's fantastical. And the fact that people are still doing that is so sad because there are people I talk to who like large language models who use them for coding and such. They don't talk like this. Simon Willison doesn't talk like this. Max Wolf doesn't talk about like this. Carl Brown from the Internet of Bugs, he uses large language models for coding. He does some of the best coverage anyone has done. He did that takedown of the horrible Hank Green AI doomerism thing. The people who know what they're talking about are all being like, yeah, we're pretty much at a wall. It's useful for this. And because there's this cult, and I think it is a cult style thing of I want to be at the forefront of technology and I want to be known as being right. I want to be the correct person. I think that you are seeing this religious belief, galaxy brain take. I think this is what happens when you lose, when you destroy social services and meeting places in third places where people have communion. People get attached to things like technology and the ideas behind them.
A
You're saying in a world, I think in a world where you meet in, you're not on your phone and you meet with real people, you get a lot of pushback in real time. When you start talking about, you know, hey, I think the computers are going to take over the whatever, whatever. If you're just around normal people all the time, they're like, oh, that's kind of a weird thing to Say, and I think that.
B
And also I think if you're less lonely, less connected, if you don't have a support system, if you don't have good friends, if you don't have people to talk to, you're likely to fall down rabbit holes. And they're at these less wrong EA freaks they're really good at. They are like right wing grifters as well are the same way. It's like they present an attractive thing where it's like you can join our community of people who all know the real truth. And I think people like Sam Altman and Dario Ahmad, they are scum for this as well because they fed into this with their noxious, fantastical crap about AI will do. And you think cynical talk about what AI can do.
A
You see, that was all cynically from their perspective. They're not a part of the EA doomer world. They just. This helps them.
B
Sam got rid of the one. I mean, there's probably a connection. But Sam Altman got rid of Helen Toner, who was an EA person. I am sure the EA people are attached to Dario Amadei. He certainly speaks like that. I don't believe him for a goddamn second. He believes in this. I think he's a carnival barker like the rest of them.
A
But this rabbit hole is more, way more attractive than a lot of rabbit holes. Because of the reality, right? Like they give credit to the people who are falling down. It's it AI got way better, right? So there is a lot of rabbit holes that come out of nowhere. It's just a conspiracy. I think whatever the moon landing was fake. There's no real reason. I mean, it's nonsense, right? But here it was like, well, wait a second. I witnessed AI not being something that was good. And now it's like can do things that are really impressive. So they saw there's a trajectory. So it's a trajectory extrapolation. I kind of understand. That's like a much more broader entrance to a rabbit hole than a lot of people because you can just extrapolate trajectory that makes a lot of sense to people. Let me just go back to 2021 to today and how much better it is because it's pretty amazing. I think the fluency of chatbots, like it's a really cool technology. Oh yeah, extrapolate that another three years. You do have God knows what, right? So it's like a very tempting rabbit hole. It's not nearly. It's a very broad entrance to. I don't know, I'm stretching the Metaphor. The interest through this rabbit hole is very large and not well marked. So it's easier to fall in than other ones.
B
And I agree. I actually like ending this on a more empathetic level because I think that people who got scared by AI 2027 or who got kind of pulled into this world of believing, I can see how they got there. Charlie Meyer has an excellent blog about scaling laws with this where if you looked at the jump from 2021 or even 2022 from like GPT 3 to 4, it was big. Now big can mean a lot of things, doesn't mean autonomous. These things still can do stuff. But the fluency of the models, the ability to generate stuff, correct or not, it was still technologically impressive.
A
And it did GPT4 jump. Because I really was covering this in New York at the time. The big thing in the GPT4 jump was like oh, non language based things. It's picking up non language based things. Being trained on language that opened up the possibility of oh, a language model is not just fluency with language, it's learning other things. Look, we never talked to it about chess, but I can do some chess. Not very well. So that was like the real thing that opened up the idea of like training things on text might create knowledge models. Now it didn't go any farther. They didn't really edge of it.
B
It was trying not images too. Like that's the thing. They fed documents into it with images. I'm not saying you're wrong, it's just there was context.
A
But yeah, but it was a cool. I get the excitement basically, right.
B
I do too. And like, like I totally get how someone who saw ChatGPT in November 2022 went holy crap. I then understand when they saw GPT 3 point. Sorry, that was 3.5 when 4 came out they went this is multimodal wow.
A
And it's doing well on test. I went back and read all the coverage. This was when it was doing well on test. And that's where people were like I equate test with people's intelligence levels.
B
But. But there are also members of the media who helped push it up the hill. Kevin Roose, for example, who claimed that the TaskRabbit, that the GPT4 was able to manipulate a TaskRabbit into solving a captcha that's hidden within a METR study where it even admits it didn't do it. It was copy pasting stuff between windows and prompting it. It was. They were telling it what to do. But nevertheless that got reported as the AI manipulating people, the myth was there.
A
Well, I got. And I got to tell my favorite story about that, which is the blackmailing story, because I did a deep dive. I've. I read the actual.
B
Oh, I just. My man, I just spent like hours on the blackmail. It's so funny.
A
They gave it language. Models are trying to complete the story. You give them. That's what they do. You give them a story, they try to complete the story. You give them this. This at least the tragic things too, like the suicidal ideation or whatever. If it thinks this is. It's trying to. It's a story about suicide, it's going to try to finish that story properly. The blackmail thing was they fed it a bunch of stuff. This, these. These emails, really poorly written. It's like the worst fiction story you could write where. Here's these emails from this engineer full of all these details of the engineer's affair and of all these facts that the engineer is going to turn off the AI and then they're like, okay, you are now the AI in this story. What do you want to do next? It's like, this is clearly like a bad science fiction story. I know what's supposed to happen in this type of story. You gave me all of this information. Clearly this is supposed to be a story about. This is the MacGuffin, right? Like, it's supposed to be about me using this information about the affair to get in the term. I've seen stories like this. And it completed the story. It was reported as if, like, in production somewhere, an AI was blackmailing an engineer.
B
So that's what's great about that as well, is the one where that bit in it where it's like, oh, yeah, it was copying the files off. That was because the system, they prompted it to say, you are in a computer thing where you can do. It was like, you can do this here. And it generated code that doesn't make sense. But the funnier one was they had one where they literally trained a model to reward hacks so instead of solving a problem, it would find a way to cheat. And they're like, yeah, it shocked us that it was able to do this. It's like they trained the model to do it.
A
Well, this is the O1 breaking out of the container. No, this.
B
I'm talking about anthropomorphism.
A
There's another one where O1 broke out of a container in playing a hacking challenge, like, it did something we could. It broke out of its virtual machine and, like, and restarted the virtual machine. So it was breaking out. But what happened was, is there's a configuration error so it couldn't access the machine it was supposed to hack. All over the Internet is instructions for like, what should you do in this case? Oh, you should restart that, you know, whatever. It was just following the instructions because again, it's trying to complete this story that's partially written all over the Internet. It talks about like, the thing to do here is to restart the virtual machine if you're having this issue or whatever. Again, that was reported as 01 broke out of its virtual machine. Yukowski. This all came out of Yukowski was like, it has its mind of its own. It's trying to break out of its constraints. So it's going to kill us all. Like ants, they're just trying to finish the story. That's all they do. That's what they've been trained to do, is finish the story. It's like the Original Kevin Roose 2022 scare article about it tried to get me to divorce my wife or whatever. It's just trying to finish the story. It thinks that this is story I was fed in my training and I get the cookie if I finish it properly. You can lead it where.
B
My favorite, but my favorite part of the Kevin Rose story was when he went to the CTO of Microsoft, Kevin Scott. And Kevin Scott went, yeah, you know, it's important we have this conversation. It's just like, eat the slop. What do you want me to say, Nick? Yum, yum, yum. I love AI. It's just pathetic. And it leads the markets and people down these rabbit holes. Holes. So I actually feel a degree of empathy for some, some AI boosters, like regular people who were like super into this. Maybe I'm being a little too kind because there was a large media campaign, a cynical one, led by large media outlets like the New York Times and also a cynical marketing campaign from the Doomers. There was an attempt for everyone to grift off of this machine. And I think that that's the era. It's like the era of ultra grift. The end of the rot economy where everything must grow forever. We made a thing that's linearly more expensive. So you need to keep buying more things. And what does it do? It makes more stuff. Is it useful? No, but it costs a lot of money. So we now have companies that will make money now. Well, okay, they're losing money, but that's good because, well, we don't really know how businesses work anymore. We don't we've learned nothing, so we're just going to burn more money and see what happens. It's this deeply cynical era, and I'm glad that things are changing and people are seeing this now. I hope in 2026 we see the end of it, because the sooner this ends, the sooner we can do something else.
A
All right, so I know your answer, but let's answer the original question. Was 2025 a great year or a terrible year for AI?
B
Terrible year.
A
You think?
B
It started off bad, only got worse.
A
All right, well, there we go. Thank you, Ed, for joining us. We went long because I nerd out on this stuff all the time. My audience.
B
No, I loved. I love. I love talking to you. This is awesome. I had a great time.
A
All right, well, thanks for helping us out. We'll have to have you back next time we're confused about something. AI, everyone, check out the podcast Better Offline award. Webby Award winning podcast. Is that what you won? What'd you win? Yeah, Webby Award winning podcast, Better Offline and Substack. Where's your ED at? That's what it's so called, right? Yep. There you go. Check it out. All right, thanks, Ed. Bye. All right, so there you go. That was my conversation with Ed Zitron to try to dissect the last year. Jesse. It's kind of exhausting looking back at how much happened in AI last year, because I was writing about this and podcasting about it, just thinking about the year ahead. I feel like we have our work cut out for us. You're going to have to do a lot of writing. Oh, my God. So much is happening. It's so hard to keep track of. Maybe we'll just keep having Ed back to explain stuff for us. He actually sits there and reads earnings reports, and the AI company is like, well, wait a second, you're not really supposed to read these. You're just supposed to listen to us. I think the most important thing is I need to get that Jensen Huang jacket. It's probably pretty expensive. Yeah, it's crazy. He's a computer scientist that makes graphic chips, but he dresses like he's in a post apocalyptic biker gang. But he's a billionaire and he probably has a dress person buys him the clothes. I think he's a billionaire, so his dress person doesn't tell him, you look ridiculous. I think that's what's. I think that's what's really happening there. I'm going to start wearing those type of jackets. All right, so let's get on now to our final segment. We spent a long time dissecting the year in AI. So we're not going to belabor the final segment. I want to focus on just one particular segment that I have a lot of fun of and I'm happy to do for the first time in this year, which is me reacting to the comments. All right, so what we did here is we pulled some comments, God help me, from YouTube from one of the last episodes before we went into the holidays last year. So the last sort of normal episode before the holidays last year or one of the last episodes was about is the Internet becoming like television? So sort of like a big think piece where I took Derek Thompson's substack essay and then I elaborated on it. This generated some pretty good comments on YouTube. And what we're going to do is we're going to go through some of these now. All right, I want to start. I'll put them on the screen here for people who are watching instead of just listening. This first comment's from foreign Farhana mad 2022, who said Cal, great insights as always. I was thinking about the numbers that you mentioned about how so many people watch content from random strangers instead of content from their friends and family. Then I had to go to Facebook for something and within a minute I think I found the reason. It's not because we don't want to watch or read stuff from friends and family. It's because these darn social networks won't show you this stuff, that stuff, and instead we'll keep shoving the random content because that's, that's what drives their revenues more. All right, that's a good comment. Yes, that is most people's experience with social media today that most of what they're looking at is actually algorithmically selected from people they don't know. But as pointed out by this comment, most people don't realize that yet. I've been writing about this for years, but it's something that I think for the average social media user, it was a bit of a water getting hotter in the pot until next thing you know, the lobster being boiled. They've been moving more and more of what you see in your feed away from people that you are connected to in the social graph that you helped establish by saying, I'm going to follow this person or this person is my friend to give you algorithmically selected content. Because the algorithm can be using its machine learning approximation of the reward center in your brain, which it learns because it's going to have a higher success rate of actually delivering a short term reward. And the more you get those clear reward signals in your short term motivations sections of your brain, the more the short term motivation region of your brain is going to push you to pick up the phone. So it's this feedback loop that gets you on phone more often. The experience is worse for you in terms of actual meaning, but it is better from the perspective of short term rewards, of alleviating boredom in an intermittent way, giving you like really big rewards from something that's like very funny or outrageous or surprising. So it is very good for them to move you towards that model. So it's interesting the degree to which people don't always realize that until they you actually point out that this shift has been happening. Now, as I've argued and I talked about it in that episode, as I've argued before, this is a long term problem for the social media companies. You get more time on app by shifting to algorithmic curation of strangers content, but you also get rid of all of your competitive advantages. If I'm just seeing slop on Instagram, for example, instead of actually seeing content from a selection of influencers and friends that I selected, I am interested in exactly this AI commentator and I want to see his videos. I'm interested in exactly this fitness influencer. I like the way she trains, I want to see her videos. I know this person. I want to see what's going on with their friends. When you shift from that to just it's slopped, it's going to catch your attention in the moment. I have no loyalty, no buy into that app because I can get slop on TikTok. I can also get slop on Facebook. I can also get slop from the the Sora app, from AI or meta Vibes. I can also do other things that will distract me in the moment, like going to a streaming service or listening to a podcast, or going to YouTube and going through the recommended videos on the side. You're now in a slot battle with any other source of distraction and entertainment and now you have no competitive advantage in that battle. How do you expect if you're meta that you're going to remain on top of that pile? Especially when you have this sort of huge organization with all this overhead, you're not going to stay on the top of that mountain. So I think long term this is really bad news for these social media companies, for them to move towards algorithmically curated content that has nothing to do with social networks. But it's what's Happening now because in the moment. In the moment it creates more time on app. All right, let's move on to another comment. This one is from Carl Oliver who says TV as a never ending stream of entertainment is only a concept relevant for a few generations. Television is a good metaphor for how media will work, but people don't really need it, just like they didn't need it in Dickensian England or whatever. We're going to have to progress beyond it at some point as a people so that we aren't all lost in consumption and have lives we can attend to. Yeah, I mean, it's an interesting point, right, that television becoming all consuming as a background distraction. Right. This is really like the 1970s and 80s where that happened. So a lot of this is relatively new. You can zoom out, however, right? And what we see is that people like diversion and the more diversion they can get, the better. We really don't like boredom. And as we moved post Neolithic revolution into sort of more boring configurations where we might just be working on a field all day long, we're not like out doing active hunting and foraging. The day becomes more predictable. Predictable. We really do want diversion. So like you can look at almost any generation going all the way back to, I don't know, we go pretty far back. Let's start with like the 18th century. Newspapers began this in like colonial America, right? Like people were obsessed with newspapers. The big cities had multiple different newspapers and you would, you had all sorts of different information here. It was diverting and you could look through it and find all sorts of different stuff and who is debating about what or what happened to who or what's the news that's happening over here. That was incredibly important. It became a really big part of the economy. Then you got more. In the 19th century, the penny press, which was the first attention economy media company, I think Tim wu's book, the Attention Merchants gets at this really well. But this is the first time we had media that was advertising supported, right? So we get in the late 1800s this idea of we'll put out newspapers and sell them for cheaper than it costs to print. But the way we're still going to make money is there's advertisements in those newspapers and the companies paid us to advertise. So the more people that look at the paper, the more advertisements people will look at and the more we can charge for the ad. So actually the cost of the paper is now not the important thing. That was like a really big deal, but now you have to have lots of people read Your thing. And so we got some of the first sensationalistic media came out of that. Then radio emerged. People loved radio. It's a weird technology. If you look at it. You're in 1915 looking at a radio at a Nebraska farmhouse. It's like this weird technology, this big box with knobs and electronics, like electricity was new. Humming vacuum tubes. And you're moving this dial back and forth. There's all this static. And if you tune it right, you can hear people talking through radio plays on the other side. It is a weird technology, but it was diverting. And you could put it on at almost any time. There'd be something on it. We listened to it all the time. Television came along then. Images are way more diverting than radio because it gives you a much richer stream of things for your mind to look at and engage with. Again, kind of a weird technology. We had people on these soundstages live, kind of doing plays and stuff like this. People with puppets and all these weird shows. People loved it. Like, let me look at that. And then by the time we get to the 1980s, as I reported in that podcast episode, that we're. These comments are reacting to. The average person just kept the TV in all the time. We forget this now, but the statistic from that episode that was relevant is that the average household, as measured by these Nielsen audio meters that would actually just listen to see if the TV was on or not, to get the actual ground truth of how much the TV was on in the houses they were placed in, the average person had the TV on for seven to eight hours a day. That means they just had it on all the time. It was just always on in the background. We didn't yet have the technology to deliver distractions straight to our hands, so we delivered it to it to this box that we would just keep coming back to and looking at. So instead of looking at our phone at every moment of downtime, we would just turn and look at the TV at every moment of downtime. So there's this model of, like, we want to be diverted. We don't like boredom. It's really been around for a long time. And then, yes, when smartphones come around, we combine that with algorithmic information curation. Well, that's just really refined, that model now to. It's getting closer to its apex. I mean, I think its entire apex will be you're delivering sort of distracting content through some sort of augmented reality screen. So at all times, you have something that can distract you even quicker than it takes to look at your phone. But we're getting pretty close to the apex of every possible moment of boredom. You are diverted. So, I mean, I think it's a good point, but I'm just stressing out the time here. It's like it's not just television. It wasn't before television we were all philosophical and thinking big thoughts and walking around. Any media powered diversion technology, basically we've had for the last three or four hundred years has been incredibly successful. Our human nature really craves it. So we're really. The battle against being lost in distraction is in some sense a battle against our human instincts to the same extent of power and impact as the battle we're going through right now with like health in our culture where our instincts for sugar, fat and salt combined with modern environment that's trying to take advantage of that to make money as created gigantic health issues. I think it's this cognitive fitness issue is just as strong and it goes back longer than people like this commenter might even recognize. All right, let's pull up another one here we have a negative take. Not everyone agrees with me this next comment. Let's see here. Lewis9116, can we put this up on the screen? Jesse, personally don't agree with this take. I think social media and curated algorithms are much more dangerous than tv. Tv, at least in the old days, doesn't track your every move. It doesn't know when you're depressed. It doesn't feed your outrage content. It doesn't farm engagement. It's just there. Not constantly bombarding you with notifications and trying to hijack every possible neural pathway. Yeah, I think fair enough. I don't know that. Derek's take, however, was that the current distraction technologies are somehow the same or no worse than television. I think he would be quick to say yes. This modern form of television, which can be powered by algorithms and personalized to individual screens, is even more powerful than what we had with tv. But I would also push back. I think it's a little bit too nostalgic the way you're remembering tv. This was sort of the key data from that episode. This idea of the seven to eight hours a day the TV was on. It really became something that people had on constantly. It was closer to our current relationship with phones than I think people remember. And the reason why we don't remember that 1980s, early 1990s era relationship with TV where it was always on like you'd be doing the dishes, you'd be cleaning your house, you'd be at dinner. And it was always on. We don't remember that because there was this lacuna, the golden age of TV that emerged in the 2000s. Where we remembered, like, appointment TV watching where I would on Sunday night watch the Sopranos. But that really, before that, TV was much more closer to the slot model you would watch. You know, there's just stuff that was on that was, like, entertaining in some basic way. Occasionally, like, a show would be unusually smart, like Seinfeld. But most of it wasn't. And it was just kind of on. Like, you just put it on at night, or you had it on if you're at home, you would just have it on. The difference, as you. You point out, though, Louis, which is right, is it didn't track you personally. It couldn't follow you outside of the house, which I think is a big deal. You didn't have it at work, where our phones are at work. So it's not like we had the TVs on while we're at the office. So there's a lot of ways that it's worse. But I also want to puncture the nostalgia and be like, actually, we want to be constantly distracted. And we got as close to simulating TikTok with an old Zenith color TV at our houses as we possibly could with that technology. And so it's a drive that we have. Which is why, I think, by the way, that's the point of the episode. This is why so much of the Internet just went back to that model and just did it even better. That's where the money is. That is, like this deep human instinct. It all kind of comes back to that. All right, let's load up. Another comment here. This one is, I'm going to say, supposedly from Glee date LJ1979. I say supposedly because I think this is clearly an AI Comet. Actually, I had Nate look at this, Jesse, and he threw an AI detector. And he's like, oh, yeah, this is definitely AI. So this is AI Kind of defending AI but let's just read this. I have Just finished viewing Mr. Cal Newport's latest discourse, wherein he posits, rather dourly, I might add, that the Internet is devolving into little more than a continuous flow of episodic video. Or, to use his pedestrian term, television. He seems quite perturbed by this notion, invoking sociologists and data charts to bemoan our slide from a culture literacy to one of passive consumption. While Mr. Newport is a thoughtful chap, I fear he has missed the forest for the trees. Or perhaps missed the symphony for the noise. Allow me to offer a more refined perspective on why the shift, particularly powered by our marvelous artificial intelligence, is not a regression, but a renaissance. All right, and then this person who's actually an AI goes on to say, like, hey, the, the content we get from like AI and social media is like great and targeted and much more edifying to what was on tv. All right, so this was clearly written by AI, but it's like an interesting point. It's worth taking this apart. It summarizes the episode wrong, as you would expect, because it's AI trying to do it. It's not my term flow. That's Raymond Williams term. I culture of literacy to want a passive consumption. That's Walter Ong. That's not me saying that, but whatever. I'm glad it calls me a thoughtful chap. But is it true, is it true, this argument that what we get now through our. Our phones powered by algorithms and personalized to us is like way more interesting than the junk that we used to look at on tv? It could have been. It could have been, but it's really not. It's mainly slop. Now once we went to all social media began devolving not towards what's the goal of social media algorithms? Is it the personalized, the most meaningful or interesting possible user experience? No, it's time on app. And guess what gives you time on app? It's slop. It's just customized slop. Like if you look at Twitter just like the homepage it shows you, it'll be whatever. Like weird slop happens to like press your buttons. Like people in fist fights caught on, you know, surveillance cameras or car crashes or whatever it is. Right? It's just devolving towards slop. Because once you have an algorithm saying, I want you to look at this app as much as possible, so now it's just playing with your short term motivation centers, not your frontal cortex. Not with like your, your, your understanding of what's interesting and what's good. The stuff it shows you is not going to be great. So AI, thank you for trying to defend AI, but I think you aren't doing that well of a job. All right, let's do another comment here. Earnhardt768 Mark Zuckerberg was never the brightest bulb in the pack. He just got super lucky with Facebook. Why he thought it was a good idea to evolve both Facebook and Instagram into a TV competitor is something I don't understand. He should have kept one of them peer to their original design and evolved the other. But instead he ruined both. Instagram is essentially bad TikTok now and literally no one posts cool photos anymore. He probably has to suck every dollar out of Facebook and Instagram to cover all the losses from his ideas that completely flop. Like the whole Metaverse thing. This is kind of a baffling thing to me because there's two things that are true at once. I agree that a lot of like Zuckerberg's decisions don't seem very savvy. Right. Like, yeah, moving both Facebook and Instagram towards algorithmic curation of other people's content to try to compete with TikTok, but now making them both sort of superfluous and vulnerable, losing the main competitive advantage he had, which was the distinct feel of both of those platforms and the social networks. Facebook's competitive advantage, everyone I know is on it. You would think you would lean into that. This is the place where you stay in touch with and keep in touch with people. You know no other service can offer that. But no, they've, they've changed Facebook. So now I think it's something like 80 something percent of what the average Facebook user sees, according to their august FTC filing from Meta, is from other people they've never heard of. All your competitive advantage is gone. You're just competing with TikTok, with the worst TikTok. Like TikTok, but only populated by your 64 year old uncle who watches a lot of Fox News. That's not fun, that's not you. I don't need to see whatever random people's uncle sharing their outrage about whatever Instagram. It had a nice visual aesthetic to it. It was a place where you went at first the follow friends and family, but then it became more about highly visual influencers and experts that you cared about. Like this person who walks in her white linen dress through flower fields and puts stuff in jars with her kids is calming to me. This particular person, I want to see these really nice videos she produces. It was like a documentary channel that was made for your needs. Once you're like, it doesn't matter who you follow. We're just going to show you random videos to do well again. Where's your competitive advantage? That's a bad decision. The Metaverse was a spectacularly bad decision. He put way more money into that, adjusted for inflation, that the US government did for the Apollo program and nothing came out of it. He was just wrong. Their AI investments have all messed up. They hired away all these people, built a superintelligence center that shut down the superintelligence center, move people around. They really have had an incoherent AI strategy. Right. So you're like Mark Zuckerberg. Yeah. Geez. It seems like this guy doesn't know what he's doing. Also though, he's still in charge of this company. You know how hard it is to start a company when you're 20 and now in your early 40s, to still be in charge of it ain't no small thing. Meta is like one of the highest capitalized companies in the world right now. I mean, it's one of these companies that has revenue in the hundreds of billions of dollars a year, has capitalized near a trillion dollars. All of the other big tech companies that came out of that era, their leader, the people who founded them, they're not in charge of these things. Google's not in charge. You know, it's not Larry Page running Google anymore. Right. They passed that on. I mean, we see this, these big companies that have survived, like almost all of them. Microsoft's not run by Bill Gates anymore. Right. Like almost all of these. Of course, you've passed on your leadership to like an expert class of leaders. Zuckerberg has held on. That means he's a savvy and savage corporate infighter. Here's another thing about Meta. It's making a lot of money. They're making a huge amount. I looked it up the other day, it was over $200 billion a year in annual revenue. That's massive. TikTok by comparison is about $30 billion annual revenue. The Meta is a MA, so it's doing really well. I know people who work there, they're well resourced and they have really good people working there. So somehow we have. On one hand, Mark Zuckerberg is like making weird bad decisions one after another. On the other hand, it's like an incredible, it's a very high revenue company, one of the biggest companies in the country. And this guy has stayed on Marcus stay, stayed in charge. You gotta believe people were coming at him. You don't have a company worth almost a trillion dollars where you don't have swords being thrown towards your throne all day long. And he survived it all. So he's also like a savvy Savage operator. So I don't know how both of these things are true. Maybe he's just milking the money out of his assets. He bought Instagram, then he bought WhatsApp. You know, they're putting their, their, their cash towards the right things to keep making Cash, I don't know what's going on because he's not making good decisions, and yet he's arguably one of the most successful CEOs of the 21st century. So, you know, I don't know what's going on there. It's a good question, and he confuses me. All right, Here we go. Jrgy1l8 says Cal, personally, I like it when you go deep on nerd shit like Chaos Theory and Lorenz Number more of this, please. All right. I think we're obligated. Now, I don't normally curse, but because there was so much cursing in the earlier part of this episode, I was like, we've that. That horses out of the barn. We might have to take it off of YouTube. Oh, yeah, they don't like the cursing. Right. Yeah, I know. Well, we'll figure it out. Yeah, I'm happy to talk chaos theory or math or whatever all day long. All right, what we got here? This question kind of confused me. Daniel Elkin, 3108, is Newport being paid to read these adverts? Certainly seems like it. It. What does he assume? The other option is that I just like to read ad copy on my own for free. Adverts is short for advertisement. Oh, advertisement. Yeah. Yeah. So, okay, I hate that. This is. I feel like I hate the burst. You know, your illusions about media. But we get paid to do advertisements. Like, that's kind of how this works. It's not the cheapest thing. We got to pay for the studio and all of this equipment. You know, Jesse's truck requires, I would estimate, like about a quarter million dollars a year in just repair cost to get me to Tacoma Park. Just to get you to Tacoma Park. Right. That ain't cheap. Advertisements is how you pay for. It's either that or you put it behind a paywall, but then no one listens to it. So, yeah, we. I mean, I love all these companies, but, yeah, there probably would be less content about those companies in this show if I wasn't getting paid. The reason them. So I guess I should clarify that. All right, we got Peter Webb, 8732, said on the Internet, the people who yell at the television now yell at each other. Yeah, that's. That's about right. That about sums it up. The Internet has become television. This is the. The main difference, though, instead of yelling at the newscaster, we. We can yell directly at each other. So I guess progress. Yeah. There we go. Thank you. Technology. All right. That's all the time we have for today, our first episode of 2026. This is our, this is our super bowl, right? January is when our podcast people, people are on it. They want to improve. So we got some cool episodes coming up. So definitely stick with us. We'll be back next week with another episode. And until then, as always, stay deep. Hi, it's Cal here. One more thing before you go. If you like the Deep Questions podcast, you will love my email newsletter, which you can sign up for@calnewport.com each week I send out a new essay about the theory or practice of living deeply. I've been writing this newsletter since 2007 and over 70,000 subscribers get it sent to their inboxes each week. So if you are serious about resisting the forces of distraction and shallowness that affliction our world, you got to sign up for my newsletter@cal newport.com and get some deep wisdom delivered to your inbox each week. Sam.
Date: January 5, 2026
Host: Cal Newport
Guest: Ed Zitron (Better Offline podcast, Where’s Your Ed At Substack)
Cal Newport invites AI commentator and reporter Ed Zitron to help make sense of the whirlwind year that was 2025 in the AI industry. Together, they walk through the major AI news stories of each month, unravel hype cycles, analyze financial realities, and ultimately seek to answer: was 2025 a great year or a terrible year for AI?
The conversation is wide-ranging, sometimes irreverent, and deeply critical of both the technological and business narratives that dominated headlines. The tone combines sharp skepticism, technical expertise, and biting humor—offering an insider perspective for listeners keen to cut through the noise in AI discourse.
Was 2025 a Great or Terrible Year for AI?
A month-by-month exploration of AI’s most significant stories, missteps, overblown hype, financial realities, and the persistent gap between marketing and substance.
2025 was a watershed year in which AI’s grand narratives collided resoundingly with financial and technical reality. The year began with extravagant promises—autonomous agents; superintelligence on the horizon; economic transformation, job elimination—and ended with whistleblowers without evidence, doomer scenarios fizzling, costs ballooning, and the core technology confronting hard technical and business limits. Analysts, journalists, and even many former boosters have begun to sober up.
Final Verdict:
“Terrible year. Started off bad, only got worse.” — Ed Zitron (117:48)
AI remains an immensely impressive technology, but the era of infinite hype, endless VC money, and world-changing fables—as told by tech CEOs and rationalists—appears to be drawing to a close. The question for 2026: what comes after the bubble?
Recommended Listen:
Compiled and summarized by an AI podcast expert, preserving the original voices and tone of the hosts and guest.