
Ragebait, sponcon, A.I. slop — the internet of 2026 makes a lot of us nostalgic for the internet of 10 or 15 years ago. What exactly went wrong here? How did the early promise of the internet get so twisted? And what exactly is wrong here? What kinds of policies could actually make our digital lives meaningfully better? Cory Doctorow and Tim Wu have two different theories of the case, which I thought would be interesting to put in conversation together. Doctorow is a science fiction writer, an activist with the Electronic Frontier Foundation and the author of “Enshittification: Why Everything Suddenly Got Worse and What to Do About It.” Wu is a law professor who worked on technology policy in the Biden White House; his latest book is “The Age of Extraction: How Tech Platforms Conquered the Economy and Threaten Our Future Prosperity.” In this conversation, we discuss their different frameworks, and how they connect to all kinds of issues that plague the modern internet: the feeli...
Loading summary
A
It's tax season, and at Lifelock, we know you're tired of numbers, but here's a big one you need to hear. Billions. That's the amount of money and refunds the IRS has flagged for possible identity fraud. Now here's another big number. 100 million. That's how many data points LifeLock monitors every second. If your identity is stolen, we'll fix it. Guaranteed. One last big number. Save up to 40% your first year. Visit lifelock.com specialoffer for the threats you can't control. Terms apply.
B
SA.
C
When was the last year that the Internet felt good to you? I think everybody has different answers to this. Mine, I think, go fairly far back, maybe to the heyday of blogging, at least before the moment when Twitter and Facebook went algorithmic. But whatever your answer to it is, I have not found many people who think 2026. Right now, this Internet, with all of its anger and its outrage and its AI slop, this is what we were promised. This is living at the technological peak. But even if there is this growing consensus that something went wrong with the Internet somewhere and that it is driving our society somewhere we don't want it to go, there's not really a consensus of what to do about it. What to do about these giant platforms increasingly spammed up with ads and sponsored results, boosting content that will keep us hooked and angry, isolating and dividing us into ranging our politics while making a few billionaires ever richer, held up by an army of low wage workers in warehouses and on delivery bikes. Something has gone so wrong. But what do we do about it? My guests today have two theories of the case. Cory Doctorow is a longtime blogger, an activist with the Electronic Frontier foundation, and a science fiction writer. His new book is Insidification why Everything Suddenly Got Worse and what to Do About It. Tim Wu worked as a special assistant to President Biden for technology and competition policy. He's a professor at Columbia Law School and author of influential books on technology, including his latest, the Age of Extraction. How Tech Platforms Conquered the Economy and Threaten Our Future Prosperity and Shidification and Extraction. Those are the ideas I wanted to put in play together here and to also think about what solutions they might present. As always, my email is recline showy times.com. Tim Wu Cory Doctorow, welcome to the show.
D
Thank you very much.
B
Great to be here.
C
So I just learned that you both went to elementary school together.
B
Yeah, that's true.
D
In suburban Toronto, a weird little school with like 80 kids. It was kindergarten to eighth grade. And one classroom, older kids taught the younger kids. We more or less were left to go feral and design our own curriculum. They chucked us out of the school on Wednesday afternoons to take our subway pass and find somewhere fun in Toronto to go do stuff. It was great.
C
Is there anything about that school that would lead people to becoming sworn enemies of our tech overlords?
B
Well, we loved tech at the time. I mean, we were early in on Apple iis. And frankly, that's where it all started in a way. You know, both of our books have this kind of pining for a lost age. And I think some of it is this early era of computing when we were just like, so excited, so optimistic. Everything was just going to be so amazing. And that, to me, a little bit, was fifth grade or grade five, as we say, programming the Apple ii.
D
Can I slightly problematize that? So we were both also science fiction readers back then. And so, you know, 1981 first William Gibson story has been out for a couple of years. I was pretty alive to the dystopian possibilities of computers at the time. So I wouldn't call myself optimistic. I would call myself hopeful and excited, but not purely optimistic. And I'd also like to say that, like John Hodgman, nostalgia is a toxic impulse. And when I think about what I like about those days, it's not that I want to recover those days. It's more that I kind of dispute that the only thing an era in which people had lots of control over their computers could have turned into is one in which the computers had lots of control over them. That there is probably something else that we could have done.
C
When you're spending time on the Internet, what feels bad to you about it?
D
So what I would do is contrast what happens when things aren't great now with how I felt about what happened when things weren't great before. So I think when I was a larvae on the early Internet and I saw things that sucked, I would think, someone's going to fix this and maybe it could be me. And now when I see bad things on the Internet, I'm like, this is by design and it cannot be fixed because you would be violating the rules if you even tried.
C
Tim, how about you?
B
I feel it's like a tool I cannot trust. I feel like the tools I like in my life, like a hammer, I swing it, it does something predictable. The Internet seems like it's serving two masters. I search for something, I get a bunch of stuff I don't really want and I don't really know what I'm getting. I want to write one email or check one thing. I end up in some strange rabbit hole and like three hours go by and I don't know what happened. So I feel like I'm constantly at risk of being manipulated or taken from and I don't trust the tools to do what they say they're going to do. And I feel that makes using it kind of like living in a funhouse.
C
So I want to make sure I give voice to somebody who is not in the show at the moment because this has. It's going to have the flavor of the prophet.
D
Elijah has entered the chat.
B
Yeah, right. Yeah.
C
Three, three middle aged guys who think the Internet went wrong somewhere along the way. When I was working on this episode with my producer, one of the interesting tensions behind the scenes was she doesn't think the Internet is bad. She thinks TikTok is, she said, a perfect platform. She has young kids and feels Amazon is a godsend for a young parent. Obviously there are many people like this who are using these platforms freely of their own volition, happily. So what do you say to somebody who says, what are you fucking all talking about?
B
Yeah, I mean, I guess I'll start. I think that, well, the middle aged thing used to be better, which is, I don't want to fall into that sort of situation. I just think the deal is not what it could be. And I think that maybe as a consumer who sort of lightly uses this, the Internet is still useful. But if people, I mean, I have children too. And I think it's hard to deny that social media has been tough on kids and has had all kinds of negative effects on that. And that really started accelerating over the last 15 years or so. I think we have a highly polarized political structure which is made worse by social media. I think we have a problem with inequality, which has gotten worse and worse and is accentuated by the fact that the margins are just so thin for independent business. And I also think this vision that it would be this equalizer leveler, this technology that made a lot of people rich, not just a few people rich, rich, that it was more or less reasonable and a lucrative thing to do to start your own business, that it would sort of change some of the challenges of inequality and class structure in the United States. Now, maybe those were very high hopes, but this is the key concept in my book and I think key to understanding the economics of our time is the importance of platforms, which are any space or any institution that brings together buyers or sellers, speakers or listeners. Every civilization has had platforms. I was in Rome a few weeks ago, and you go to the Roman forum, and there it is. It's all together. The buyers, the sellers, they have, the courts they have where people gave their speeches. They're kind of the core of every civilization. And at some level, why I wrote this book is I was interested in this question of what our fundamental platforms look like and how that reflects on the civilization we are building, because I do think they have a large impact. I think that's kind of undeniable. I think that things have gotten worse in many dimensions, and I guess it relates to my view of the state of the country as well. I think we've been in a better place in other periods of American history. And I think the Internet's not the only cause, but I think it's part of it.
D
If I were having this conversation with your producer and we had some time to talk about it, I would probably walk them through a couple of the undisputed ways in which some people have found the Internet get worse for them. So Tim has talked a little about margins for small businesses. There's also people who are performers who found that the take that's being sucked out of their pay packet every month is going up and up from the platforms. There's people who would not like to be snatched by Ice Snatch squads who installed Ice Block on their iPhone, only to have Tim Cook decide that ICE officers were a member of a protected class and removed that app. And now you can't install that app because the iPhone only lets you install official apps. And I'd say, like, just because this hasn't hit you, I think unless you have a theory about why you are favored by these platforms, then you should at least be worried that this would come. And I would follow up by saying, like, let's not fall into the trap of vulgar Thatcherism. You know, Thatcher's motto was, there is no alternative. And I think tech bosses would like you to believe that too, that if you're enjoying having a conversation on Facebook with your friends, which I stipulate lots of people do, I think that's absolutely the case, and we should value and celebrate that. That you just have to. That there is no way to have a conversation with your friends that Mark Zuckerberg isn't listening in on, and that to ask for otherwise than that would be like asking for water that's not wet. It's just, like, not possible. And what I'm militating for is not, you don't like that thing you like. It's like I, I like that you like the thing you like. I want to make it good. And I also want to guard it against getting worse because, just because it hasn't happened to you yet, it would be naive to think that it would never come for you.
C
So your books are two frameworks for understanding what I would call corporate capture of the Internet. The way we went from the dream of a decentralized user controlled Internet to something that a small number of corporations really run and have enormous power over. And Tim, the term you focus on is extraction. COREY the term you focus on is inshidification. So I'd like you just both to define those terms for me. What is extraction, Tim? What is insidification?
B
Corey so extraction is actually a technical economic term that refers the ability of any entity or any firm with market power or monopoly power to take wealth or other resources far in excess of the value of the good being provided. Not only the value being provided, but also its cost provided. That's the technical definition. You know, outside of tech, it might be a pharmaceutical company, they have, you know, there's a rare disease, they have the only treatment for it. And you know, maybe they're, they're extracting as much as they can. You know, $100,000 a year is about the usual they, they pin those kind of diseases. That, and that's the definition. And I think the idea of it comes from a sense, something I get from teaching at business school sometime, is that American business has, in my view, moved increasingly to focus its efforts on trying to find points of extraction as a business model, as opposed to say, improving the product or lowering the price. Try to find the pain points where your customers really have no choice and then take as much you can. Kind of like in a poker game when you go all in because you got the good hand. Now there's always been a little bit of that in business, or maybe a lot like in the Gilded age. But the question is what is the ratio and how much of business is providing good services for good prices? You know, making a profit, that's fine. And how much is just that different thing of extraction.
C
So Tim, I want to, before I move on to Corey, zoom in on something you said there because a lot of that definition seemed to turn on how you define value. And I mean a lot of economists would say price is a method of discovering value. If you have a pharmaceutical, people are willing to pay $70,000 for that means they value it at $70,000. Even if you think that is extractive so how do you know when a price, when a profit is actually extractive versus when we're just seeing that people value that product very highly and bully on the producer for creating something people value so highly?
B
Yeah, so if someone, for example, has no choice, but they are desperate, let's say for water, and someone is able to sell them a bottle of water because they're dying for $100,000 or something like that, yes, that person does value it at that level. But we would like to have an economy, and every economist would agree with this, is that an economy full of nothing but maximize monopoly prices, where people are in a position to extract or firms is inefficient for two reasons. One is too much money gets spent on that water versus other things like maybe pursuing an education. And second, that the entity that holds that much power actually has an impulse to reduce supply, reduce output, and therefore produce less of the stuff so that they can extract the higher price. I mean, this is just classic monopoly economics. I guess I'm getting into. Everyone inside themselves has something they are willing to pay, but that doesn't mean it's a good society. When you're constantly paying the maximum you are willing to pay in every situation.
C
For many of these platforms, for a Facebook, for a TikTok, we're not paying for them. So when you say they are extractive, what are they extracting and from whom?
B
So when you use Facebook, you are constantly being mined for your time, attention and data in a way that is extraordinarily valuable and that yielded something like 67 billion in profit last year. Things that feel free. Is it free when you suddenly spend hours wandering around random things you didn't intend to? Is it free when you end up buying stuff that you didn't really want and wonder why you get it later? Is it free when you feel that you've had your vulnerabilities exploited? I would say none of that's free. You're poorer both in your own consciousness and in terms of what attention and your control over your life. And you're poorer probably in misspent money.
D
I also wanted to react to something that you were sort of fainting at Ezra, which is this idea of revealed preferences which you often hear in these discussions, right? That if you let Facebook spy on you, no matter what you say about how you feel about Facebook spying on you, you have a revealed preference. And Tim used the word power when he responded to that. And I think that, you know, if you ask the neoclassicals, they'll say, well, we like models and it's hard to model qualitative aspects like power. So we just leave them out of the model and hope that it's not an important factor. And this is how you get these incredibly bizarre conclusions. Like, if you sell your kidney to make the rent, you have a revealed preference for having one kidney. But what we actually know when we give people choices, when the state intervenes or when there's countervailing power, is that often you get a different revealed preference. You know, when Apple gave Facebook users the power to tick a box and opt out of Facebook spying, 96% of Apple users tick that box. So the argument that Facebook users don't mind being spied on, I think is blown out of the water when you actually give them a way to express preferences. And I assume the other 4% were, like, either drunk or Facebook employees. Or drunk Facebook employees. Which makes sense because I would be drunk all the time if I worked at Facebook. But I think it's hard to deny that people really don't want to be spied on if they can avoid being spied on.
C
All right, I think that's a good setup to inshidification.
D
Yeah, inshidification. It's really a label. I hung on both an observation about a characteristic pattern of how platforms go bad, but I think much more importantly, why they're going bad now. Because we didn't invent greed in the middle of the last decade. So something has changed. My thesis is that some exogenous factors have changed. So the pattern of platform decay is that platforms are first good to their end users while locking them in. That's stage one. And once they know that the users have a hard time departing when they face a collective action problem or when they have high switching costs, you can make things worse for the end users, safe in the knowledge that they are unlikely to depart in order to lure in business customers by offering them a good deal. And so far, so good. I think a lot of people would echo that, but they would stop there. They would say, oh, you're not paying for the product, so you're the product. So that this is about luring in users and then getting in. Business customers will pay for it, but that's not where it stops, because the business customers are also getting screwed, because the business customers get locked in. And this power that the platform end up with over their business customers is then expressed in stage three, where they extract from those business customers as well. They dial down the value left behind in the platform to the kind of minimum like homeopathic residue needed to keep the users locked to the platform, the businesses locked to the users, and everything else is split up among the executives and the shareholders. And that's when the platform's a pile of shit. But the more important part, as I say, is why this is happening now. Broadly, my thesis is that platforms used to face consequences when they did things that were bad for their stakeholders. And those consequences came in four forms. They had to worry about competitors, but we let them buy those. They had to worry about regulators. But when a sector is boiled down to a cartel, they find it very easy to agree on what they're going to do and make their preferences felt because they have a lot of money, because they're not competing with one another, and they capture their regulators. They had to worry about their workers because tech workers were in very scarce supply and they were very valuable. And they often really cared about their users and they could really say, no, I'm not going to that thing. I missed my mother's funeral to ship on time and make it stick because there was no one else to hire if they quit. And they were bringing a lot of value to the firm. But of course, tech workers famously thought that they were temporarily embarrassed founders. And they didn't unionize, they didn't think they were workers. So when the power of scarcity evaporated, they had not replaced it with the power of solidarity. And so now you have 500,000 tech layoffs in three years, and tech workers can't hold the line. And then finally there was new market entry. There were new companies that could exploit something that I think is exceptional about tech. I'm not a tech exceptionalist, broadly, but I'm an exceptionalist about this, which is that every program in your computer that is adverse to your interests can be neutralized with a program that is beneficial to your interests. And that means that when you create a program that is deliberately bad, you invite new market entrants to make one. That's good, right? If you lock up the printer so it won't take generic ink. You just invite someone to not only get into the generic ink business, but get into the alternative printer firmware business, which eventually could just be the I'm going to sell you your next printer business. But what we've done over 20 plus years is monotonically expand IP law until we've made most forms of reverse engineering and modification without manufacturer permission illegal. A felony. My friend Jay Freeman calls it felony contempt of business model. And as a result, you don't have to worry about market entry with this Incredible slippery, dynamic character of technology. And when you unshackle firms from these four forces of discipline, when they don't have to worry about competitors or regulators or their workforce or new market entry through interoperability, the same CEOs goes to the same giant switch on the wall on the C suite marked in shidification, and they yank it as hard as they can, as they've done every day that they've shown up for work. And instead of being gummed up, it has been lubricated by an inshytogenic policy environment that allows it to go from 0 to 100 with one pull. And that's how we end up where we are today.
C
All right, I want to bring these out of theory, though. Corey, I applaud how well structured that was on the fly and have you both walk through this with an example that you use in your books.
D
Sure.
C
And Corey, I want to start with you. Walk me through how you see insidification as having played out on Facebook itself. Not all of Meta, but Facebook, where it started, when it was adding value to users in the early days to where you feel it has gone. Now tell me your Facebook story.
D
So Facebook, really, its Big Bang is 2006. That's when they opened the platform to anyone, not just people with a Edu address from an American college. And Mark Zuckerberg needs to attract users, and his problem is that they're all using a platform called MySpace. So he pitches those users and he says, look, I know you enjoy hanging out with your friend on MySpace, but nobody should want to use a surveillance driven social media platform. Come to Facebook and we'll never spy on you. You, we'll just show you the things that you asked to see. So that's stage one. But part of stage one, remember, is that there's a lock in. It's just the collective action problem. Right. You love your friends, but they're a pain in the ass. And if the six people in your group chat can't agree on what bar to go to this Friday, you're never gonna agree on when it's time to leave Facebook or where to go next. Especially if some of you are there because that's where the people with the same rare diseases you are hanging out. And if some of you are there because that's where the people in the country you emigrated from are hanging out. And some of you are there because that's where your customers or your audience or just that's how you organize the carpool for the kids, Little League. And so we are Locked in. And so that ushers in stage two, making things worse for end users to make things better for business customers. So think about advertisers. Advertisers are told, you know, do you remember we told these rubes that we weren't going to spy on them? Obviously that was a lie. We spy on them from asshole to appetite. Give us pennies and we will target ads to them with exquisite fidelity. So the advertisers pile in. Publishers pile in too. They become locked to the platform, they become very dependent on it. And in stage three, advertisers find that ad prices have gone way up. Ad targeting fidelity has fallen through the floor. Ad fraud has exploded to levels that are almost incomprehensible. Publishers famously now have to put their whole article there, not just an excerpt. And wobitai, the publisher that has a link back to their website. Because Facebook's downranking off platform links is potentially malicious. And so they don't have any way to monetize that except through Facebook's own system. And we've got a feed that's been basically denuded of the things we've asked to see. It has the minimum calculated to keep us there. And this equilibrium is what Facebook wants, but it's very brittle because the difference between I hate Facebook and I can't seem to stop coming here and I hate Facebook and I never coming back, it can be disrupted by something as simple as a live stream mass shooting. And then users bowl for the exits. The street gets nervous, the stock price starts to wobble, the founders panic. Although being technical people, they call it pivoting. And you know, one day Mark Zuckerberg like arises from his sarcophagus and says, harken unto me brothers and sisters, for I've had a vision. I know I told you that the future would consist of arguing with your most racist uncle using this primitive text interface that I invented so I could non consensually rate the fuckability of Harvard undergraduates. But actually I'm going to transform you and everyone you love into a legless, sexless, low polygon, heavily surveilled cartoon character so that I can imprison you in a virtual world. I stole from a 25 year old comedic dystopian cyberpunk novel that I call the Metaverse. And that's the final stage. That's the giant pile of shed.
B
All right, Corey, you got a good.
C
Rant there, my man. I give you real rant.
B
Corey could be a rapper if he decided to get into that line of business.
D
The world is crying out for a middle aged technology critic rapper.
C
All right, let me ask you at least one question here. So I'm not just too taken in by your charisma, which is to say, I think that the counterargument somebody would offer is I think two things. One is for all the pivots, all the scams, by the way, I mean, I was a publisher during the era of the Facebook fire hose to publishers and the era of pivot to video, when Facebook videos were getting these astonishing view counts. And one, they kept first all the money they promised everybody come get this huge scale. We're giving you all this traffic, you can build a business here. There was no business to build there at any significant scale. And two, turned out that the video view counts were fraudulent. And so a huge amount of the news industry, among other things, pivoted to video, and it was based on lies. And there was a recent Reuters report that Facebook was actually charging advertisers more for these things that they knew were scams.
D
10% of their ad revenue is ads for scams by their own internal accounting.
C
So I am really not here to defend Facebook as an actor. But one of the crazy things amidst all of this, a thing you really focused on there was moving from showing us what we had asked to see to showing us what I would say Facebook wants us to see. There's just the FTC versus meta case. Tim was of course involved in that. And one of the statistics that came out during it is that only 7% of time spent on Instagram is spent on things people you follow are showing you. Similarly, on Facebook itself, it's under 20%. I forget the exact number, but it's very low. They have moved under competition from TikTok specifically, although not only to these AI driven algorithmic feeds showing you not what you have asked to see, but what they find will keep you there. And what they are finding is that it will in fact keep you there. And people are coming back to it and they spend more time on Instagram when you turn the feed into this algorithmic feed, this is the whole revealed preference thing that you were talking about earlier. My personal experience of Instagram, when I go on it now is one reason I try to go on it less, is that I can actually feel how much more compelling it is. I like it less, but the feeling of getting pulled into something is much stronger. And so I think if you had Mark Zuckerberg arisen from his sarcophagus. I was gonna say office, because I'm a More polite person here. He would say we did this under competitive pressure. TikTok was eating our lunch, we stole a bunch of things from TikTok and now we're doing better. We also stole a bunch of things from Snapchat and now we're doing better. Because in fact we are under a lot of competition and we are incred good at responding to that competition in ways that our user base responds to. This is not inshinification, this is the magic of competition itself. And you know that because actually look at our profit margin and look at how much we've changed.
D
So let me say that I don't think competition is a good unto itself. And I think it is absolutely possible to compete to become the world's most efficient human rights violator. The reason I like competition is because it makes firms into a rabble instead of a cartel. So in 2022, two teenagers reverse engineered Instagram and they made an app called OG app. And the way OG app worked is you gave it your login and password, it pretended to be you and logged into Instagram, it grabbed the session key, it grabbed everything in your Instagram feed, it discarded the ads, it discarded the suggestions, it discarded all of the stuff that wasn't a chronological feedback of the people you followed that they posted. Recently, Facebook or Meta sent a letter to Apple and Google who obliged them by removing the app because there's honor among thieves. So if you want to find out what people actually prefer, you have to have a market in which people who disagree with the consensus that people are kind of gut flora for immortal colony organisms we call limited liability corporations, and that they are entitled to dignity and moral consideration as beings unto themselves, those people have to be offering some of the alternatives to find out what they want. But because under modern IP law, something called the Digital Millennium Copyright act, it is a felony to modify the app without permission. When Meta sent the letter to Apple and Google, they agreed that they would side with Meta. And because you can't modify those platforms to accept apps that haven't run through the store, that was the end of the road for ogi.
C
But I think this is a little bit of a narrowed example. As somebody who gets a huge number of press releases for all these pro social apps that are built to compete with Instagram and TikTok and all of them apps that are meant to respect your attention, apps that are meant to be virtuous in a way these apps are not. And watches one after another after another after another basically go nowhere, get outcompeted. The Point I'm making is the example you are giving. They were able to basically say there was a term of service violation. Maybe they should not be allowed to do that. But there are lots of things that emerge and are meant to be better or different or something and people do. This is where I want to make sure my producer has a voice. There are people who just absolutely like TikTok. They're people who like Instagram. They know there are other things out there and they're not clamoring for a competitor or an alternative. I think suggesting that there is no capacity to switch is going a little far.
D
No, I'm not saying there's no capacity to switch. I'm saying the higher the switching costs are, the lower the likelihood that people will leave. You know when we had pop up ads in our browsers and real pop up ads, the paleolithic pop up ad, that was a whole new browser window that spawned one pixel squared, autoplayed audio ran away from your cursor. The way that we got rid of that was it was legal to modify browsers to have pop up blockers. More than 50% of us have installed an ad blocker in our browser. Doc Searls calls it the largest consumer boycott in human history. And as a result, well, there is some moderation upon the invasiveness of what a browser does to you that is in marked contrast with apps. Because reverse engineering an app because it's not an open platform is illegal under American copyright law. It violates section 1201 of the Digital Millennium Copyright Act. And so when we talk about how these platforms have competed their way into toxicity, we are excluding a form of competition that we have made illegal. For example ad blockers, for example privacy blockers, for example things that discard algorithmic suggestions and so on. Taking those off the table means that the only competitors you get are firms that are capable of doing a sort of holess bolus replacement to convince you that no, you don't want to use Instagram anymore, you want to use TikTok instead, as opposed to you'd like to use TikTok or Instagram rather, but in a slightly different way that defends your interests against the firm's interests. And we mustn't ever forget that within digital technology and living memory, we had a mode of competition that we prohibited that often served as a very rapid response to specifically the thing you're worried about here. I have a friend, Andrea Downing, who has the gene for breast cancer and she's part of a breast cancer previver group that was courted by Facebook in the early 2010s and they move there. And this group is hugely consequential to them because if you have the breast cancer gene, you are deciding whether to have your breasts removed, your ovaries removed, the women in your life, your daughters, your sisters, your mothers, they're dying or sick and you're making care decisions. This group is hugely important. And Andrea discovered that you could enumerate the full membership of any Facebook group whether or not you were a member of it. This is hugely important to her friends there. She reported it to Facebook. Facebook said, that's a feature, not a bug. We're going to keep it. They sued. It was non consensually settled when the FTC settled all the privacy claims. And they are still there because they cannot overcome the collective action problem that it takes to leave now. They will eventually, when Facebook is terrible enough, that community will shatter and maybe it will never reform. That is not a good outcome.
B
LifeLock how can I help?
E
The IRS said I filed my return, but I haven't.
A
One in four taxpaying Americans has paid the price of identity fraud.
B
What do I do? My refund though.
D
I'm freaking out.
B
Don't worry, I can fix this.
A
LifeLock fixes identity theft, guaranteed and gets your money back with up to $3 million in coverage.
E
I'm so relieved.
C
No problem.
B
Problem. I'll be with you every step of the way.
A
One in four was a fraud. Paying American. Not anymore. Save up to 40% your first year. Visit lifelock.com Special offer terms APPLY I'm Judson Jones.
E
I'm a reporter and meteorologist at the New York Times. For about two decades I've been covering extreme weather, which is getting worse because of climate change. And it's becoming more important to get timely and and accurate weather information. That's why we send these customized newsletters letting you know up to three days in advance about extreme weather that could impact you or a place you care about. At the Times, you can be confident that everything we publish is based off the most accurate, scientific embedded information available to us because we want you to be able to make real time decisions about how to go about your life. This is the kind of work that makes subscribing to the New York Times so valuable and it's how you can support fact based independent journalism. So if you'd like to subscribe, go to nytimes.com subscribe.
C
All right, Tim, I want to go to your story here. One of the core tales you tell in Your book is about Amazon. So walk me through the process of moving a platform from a kind of healthy, constructive platform to becoming an extractive platform through your kind of story of what happened with Amazon.
B
So back in the 90s, when people were thinking about what is going to be great about the Internet, there was this great, really serious promise that it was going to make a lot of people rich. It was going to distribute wealth in new ways. And when people talked about how that was going to happen, a lot of it was like, well, you know, everyone's going to be able to have their own store, sell stuff, and people are going to make a lot of money that way. So that was a big promise and I think was an important promise. Back in the day, it was mostly ebay that people were talking about. So then you have Amazon. Amazon, as you may remember, was once upon a time a bookstore.
C
I do remember that, actually. That's how old I am.
B
And you know, their basic idea was be bigger and sell more stuff. At some point they opened the marketplace, the Amazon Marketplace, which was different because it was a platform. In other words, it was a place that people could come and sell their stuff. At first it was used books, then it spread into other markets and they realized a few things. One is that fulfillment would be very important. Ebay in the old days, the sellers had to wrap it themselves and send it off. So that wasn't a very scalable model. And they had a good search engine. Amazon invested hard in search and it worked. And more and more sellers came, more and more buyers came. And so the Amazon Marketplace took over ebay and became very successful. And at that point, I would say maybe around 2010 or something like that was fulfilling what I would call the dream of the Internet age, which is a lot of people will be able to go in this place, start their thing, make a lot of money. It coincides with the rise of the blog and small online magazines. You know, that whole era that we are talking about, during that period, Amazon's take was below 20%. It kind of depends how you count, but you know, somewhere between 15, 20%.
C
Their take of what a small business is selling.
B
Of the sales. Yeah, yeah, yeah. So if you sold like a hundred dollars, they take $20. I mean, it depend a little bit. There were some storage fees and so on. So it was a good place to make money. What changed, I think was once Amazon had confidence that it had its sellers and it had its buyers more or less locked up. This is basically over the 2010s. They bought a Couple companies that were potential threats to them. Diapers.com, for example. It might seem ridiculous, but diapers could have been a way in to threaten them.
C
Why don't you tell the Diapers.com story for a minute? It's a kind of famous story in Amazon, but I think it's worth telling.
B
So there was the platform launched to be an alternative to Amazon, and their thought was every parent needs diapers delivered quickly. So why don't we make that the beginning in the same way Amazon started with books and then Amazon saw this, thought it was kind of threatening, and in the strategy of the day, just bought them. Of course, the founder's pretty happy and Amazon managed basically to capture this market. And that's when I think it turned to the extraction phase. In the last 10 years, Amazon's strategy has just basically been for its marketplace to turn the screws and increase the fees, change the margins, so that many sellers are paying over 50% or more, basically the same as brick and mortar businesses. And Amazon prices are rarely any lower. They actually have done a lot to try to prevent anyone pricing lower. And I think the one thing I would focus on is what they call advertising, which may be familiar to you as sort of the sponsored results that you get when you're searching. So what's going on there is that sellers are bidding against each other, bidding down their own margins to get higher up in the search the results. And that little trick, that sort of one weird trick has become this extraordinary cash cow. It's more profitable than Amazon Web Services, which is sort of surprising. Last year it was 56 billion.
C
Just paying Amazon for higher rankings in their search results was $56 billion.
B
Yes, 56 billion. It's looking like it's going to be over $70 billion, Corey.
C
When I'm searching on Amazon and I see that Amazon's choice looks like a little prize like that that product won a competition where a bunch of editors chose it as the best one. What am I looking at there?
D
So that is broadly part of this thing Tim was discussing, where they're piling on junk fees for the right to be at the top of the results. Sometimes it's more subtle, where if you're not paying for prime and paying for fulfillment by Amazon and paying for all these other things, you aren't eligible. And the more of these you buy, the greater chance you have of being chosen.
C
But is that people be. Are they literally paying to be Amazon's top choice? I mean, as a dumb consumer, maybe I look at that and I think, oh, this is some algorithmic combination of is it the bestseller, what are its reviews, et cetera.
D
So you're right that it is algorithmic. But the algorithmic inputs are not grounded primarily in things like quality or customer satisfaction. They're grounded in how many different ways you've made your business dependent on Amazon in such a way that every dollar you make is having more and more of that dollar extracted by Amazon. There's some good empirical work on this from Maria Mazzucado and Tim O'Reilly, where they calculate that the first result on an Amazon search engine results page on average is 17% more expensive than the best match for your search. So that's what you're seeing is basically the Amazon top choice is the worst choice.
C
So this really feels to me like a place where to use Corey's word things and shitified that when I go around the Internet now, when I play something in a Spotify playlist or click on a song I like and move to the radio version of Spotify, or when I search something on Google or when I search something on Amazon, these used to be very valuable services to me to search for something on Amazon and see rankings weighted by how popular the product is, how high the reviews are. Right. Like, I took the weighting of the search as to some degree a signal of quality. Certainly Google, the whole idea was that what comes first in search was built on page rank and it was going to be quality. Spotify had an algorithmic dimension to it, but was supposed to be showing me things that people like me would like to listen to. And now there is so much sponsored content in every one of these results, and it is so unclear what is what and who is paying for what and why I'm getting this song or that result. You know, one reason I ended up on these platforms is because I trusted these results and now I trust nothing.
B
Yeah, I mean, it's going back to the definition of extraction. I mean, we are kind of paying $70 billion collectively to make search worse.
C
So when does this move from. This is just their business model. And if you want to find something else, like go buy something on Walmart, go buy something on Target, go buy something at Best Buy. You can do all those. I've done all those. I just ordered a blender from Kohl's versus We've moved to extraction, and we should see it as a public policy problem.
B
Yeah, I think that's a really great question. It's a kind of question we faced, I think, repeatedly in history. When you start to have a business model start to settle down, you see less real disruptive competition possible. And I think at some level, once a market has settled, at some point you gotta call a limit. And we do that in many other markets. And Amazon is still a great way to find a lot of product. It's the world's largest marketplace. But I would say they're running themselves like an unregulated monopoly.
C
Both of you spend a lot of time on the number of small acquisitions that these companies make. And maybe many of them get shut down or they acquihire the top people, but they're also things that might have grown into something bigger or else on the other side. Sometimes it really is the case that a big player buying something smaller, they can scale it up into something new like Google bought. I mean this was actually a fairly big acquisition, but Waymo and kind of amazingly like they seem to have made driverless cars work. And I think access to Google's compute and other things was not insignificant in that. And you can look at other cases where these companies buying something small, they're able to build it into something that ends up being a great option in Microsoft Office or in Google Docs or whatever it might be. So how do you think about on the one hand, the ways in which, I mean, I think it's what Google's made, you know, more than a thousand acquisitions over, you know, some number of years. How do you think about the ways in which that harms competition? But also, you know, I've known founders who get acquired and are excited to get acquired because they think it will give them scale and the capacity to compete in a way they wouldn't versus Google just trying to do it itself. I think the antitrust law level is one thing, but the, the sort of anti competitive versus pro scale level is like a much bigger challenge. The way Silicon Valley now works. And I'm curious to hear you talk through the, the pros and the cons of that.
B
Yeah, I think that's a really great question. You know, Joseph schumpeter back in 1911 wrote a book about entrepreneurs basically and he said, you know, these very unusual people who are willing to go out and start a, you know, take these kind of risks, they have some vision, they do this kind of thing. And he thought they were essential to economic growth, that they were these kind of unusual, almost like superheroes and would do these things and go out and take these chances. The United States economy in general has thrived because it has a lot of those kind of individuals and they can start things. And I think we've erred too far in having all the brains under one roof. You know, it's starting to remind me of AT&T in the 60s or IBM where they sort of became much more centralized about innovation and big ideas would never be developed. It became kind of group thinky. I think when the Justice Department did a deal, sued AT&T, tried to break them up and they forced them at&t to stay out of computing forever and also license all of their patents, including the transistor patents. And all kinds of people started quitting their jobs and saying, I'm going to start a semiconductor firm. And there lies the origins of US semiconductors and also frankly US computing without AT&T. So I think we have done much better with divided technological leadership. I Frankly think that LLMs might never have gotten started without OpenAI being an alternative force because they're obviously threatening to Google's business model.
C
Although don't. In a way you have to give Google some credit on LLMs specifically. You were talking about transistors a minute ago. But Google does the fundamental research in Transformers and releases it publicly and creates a.
D
It doesn't do anything with it internally until there's a competitor that threatens them.
C
It's just striking how good of an actor they were for a period on AI specifically.
D
Right.
C
Like treating it like they had a Bell Labs.
B
I agree with that. It actually is a lot like Bell Labs in the sense that Bell Labs kept inventing stuff. I mean Bell Labs collected these a lot of amazing people and then never let things come to market. The Internet being probably the best example of it.
D
Yeah. I think when you look at these companies and their acquisitions, what you see is that these companies very quickly suffer from what both Brandeis and Tim called the curse of bigness, that they find it very hard to bring an actual product to market that they invent in house. When you look at Google, they've had like one really successful consumer facing product launch and that was in the previous millennium. And almost everything they made in this millennium failed. It either didn't launch or when after it launched, they shut it down. Whereas their giant successes, their video stack, their ad tech stack, documents, collaboration, maps, navigation, server management, mobile. Right. These are companies they acquired from someone else and operationalized. And I'm an exops guy, I'm a recovering sysadmin, so I'm not going to say that that's nothing. It is a skill unto itself. The careful work to make things work and make them resilient and Scale them. But the idea that that has to happen under one roof, I think is a false binary. I mean, one of the things Google did arguably far more efficiently than they hired innovators is they hired operations people. And those are the people who really do the yeoman service at Google. Because the innovators, the product managers, never get to launch. They only get to buy other people's products and refine them.
B
You know, it comes down to what you think of is the track record, I guess, of monopolized innovation and it has some hits. But I'm saying a much more mixed model, I think historically is a lot stronger. If you look at the entire track record of US innovation, I think monopoly innovation leads you towards AT&T, Boeing, General Motors kind of model, as opposed to what the best of Silicon Valley has been.
D
And meanwhile, I think you mentioned aqua hires for people who aren't unfortunate enough to be steeped in the business of Silicon Valley. And aqui hire is when a company is purchased not for the product it makes, but because the team who made it have proved they can make a product and then they shut down the product and they hire the team. And acqui hires are, I think, a leading indicator of pathology in tech and investment. An acqui hire is basically a post grad project where venture capitalists sink some money into you pretending that you're going to make a product. It's a science fair demo in the hopes that the company will buy you. And in lieu of a hiring bonus, we'll give you stock. And in lieu of a finder's fee, we'll give them stock. But no one's trying to actually capitalize a product or a business. I think anytime you see a preponderance of acqui hires in your economy, that should tell you that you need to sit down and figure out how to rejigger the incentives because your economy is sick.
C
Corey, we've been talking here about these markets as really having two players in them, which is, well, maybe three. We've been talking about users, sellers and platforms. But something the yearbook focuses quite a bit on is a fourth, which we need to talk about too, which is labor. There are huge numbers of people working for these companies, huge number of people delivering Amazon packages and Walmart packages. And one thing that both of you focus on is the way in which as these companies become bigger and more dominant, their labor practices can become, I don't know if insidification is a term you would use there, but shittier or more extractive. Can you talk a bit about that side of it. What has happened to the labor practices?
D
Yeah, I mean, we could talk about the other tech workers, right? The majority of tech workers drive for Uber or for Amazon or work in a warehouse. And they certainly don't get like free kombucha and massages and a surgeon who'll freeze their eggs so they can work through their fertile eggs years. They're in a factory in China with suicide nets around it. But an example that kind of pulls this all together, how you get monopoly, regulatory capture, the degradation of labor with technology that relies on blocks on interoperability. I think we could do no better than to talk about nurses. And I'm going to be making reference here to the work of Veena Duble, who's a legal scholar who coined a very important term, algorithmic wage discrimination. In America, hospitals preferentially hire nurses through apps, and they do so as contractors. So hiring contractors means that you can avoid the unionization of nurses. And when a nurse signs on to get a shift through one of these apps, the app is able to buy the nurse's credit history. And the reason for that is that the US Government has not passed a new federal consumer privacy law since 1988, when Ronald Reagan signed a law that made it illegal for video store clerks to disclose your VHS rental habits. Every other form of privacy invasion of your consumer rights is lawful under federal law. And so among the things that data brokers will sell anyone who shows up with a credit card is how much credit card debt is any other person carrying and how delinquent is it. And based on that, the nurses are charged a kind of desperation premium. The more debt they're carrying, the more overdue that debt is, the lower the wage that they're offered on the grounds that nurses who are facing economic privation and desperation will accept a lower wage to do the same job. Now, this is not a novel insight. Paying more desperate workers less money is a thing that you can find in, like, Tennessee Ernie ford Songs about 19th century coal bosses. But the difference is that if you're a 19th century coal boss who wants to figure out how much the lowest wage each coal miner you're hiring is willing to take, you have to have an army of Pinkertons that are figuring out the economic situation of every coal miner. And you have to have another army of guys in green eyeshades who are making annotations to the ledger where you're calculating their pay packet. It's just not practical. So automation makes this possible. And you have this vicious cycle where the poorer Unnurses. The poorer they become, the lower the wage they're offered. And as they accumulate more consumer debt, their wage is continuously eroded. And I think we can all understand, like, intuitively why this is unfair and why, as a nurse, you might not want it, but also, like, do you really want your catheter inserted by someone who drove Uber till midnight the night before and skipped breakfast this morning so they could make rent? This is the thing that makes everyone except one parochial interest worse off. And this is not a free floating economic proposition. This is the result of specific policy choices taken in living memory by named individuals who were warned at the time that this would be the likely outcome and who did it anyway.
C
I think this is getting at something we're starting to hear a lot about, which is anger over algorithmic pricing of various kinds. So when I was walking up to do the podcast today, that Kyron on CNN was about an investigation finding the Instacart was charging many different people many different prices. And so the price you were seeing on Instacart wasn't the price, it's your price. And I could imagine a neoclassical economist sitting in my seat right now and saying, pricing becomes more efficient when it discriminates that the market will be more efficient if it can charge Ezra a higher price for kombucha. If I'm getting that delivered because of things it knows about me and my kombucha habits, and it charges somebody else a lower price because it knows they value the kombucha less, or a nurse a higher price, or a higher wage and a lower wage depending on their situation, that in fact we're just getting better and better and better at finding the market clearing price. And this is what economics always wanted. We're finally hitting the utopia of every person having the market clearing wage and the market clearing price. Why don't you agree with that?
B
Yeah, I mean, the fundamental question is, is that really the kind of world you want to live in? In other words, do you constantly want to live in a place where you are being charged the maximum you would pay for something now, you know, that could rebound to the benefit of people who are very poor? But it is in economic terms, it's always only about producers taking everything from the market. And I just think it's a very, you know, just moving away from the. The efficiency potentially of it. I think it makes for a very unpleasant lifestyle to be constantly feeling you're being exploited. And the other thing I'll say is there's also a huge amount of effort People make trying to move what category they're in and pretend to be poor. So I think it is overrated and I guess it relies on overly simplistic models of what makes people happy.
C
There's a way in which efficiency, housing is an interesting term in economics because in economics, as in life, you want things to be somewhat efficient, but too much efficiency becomes truly inhuman. I find this even in the very modest example of personal productivity efforts. It's great to have a to do list. If I really force myself onto the scaffolding of a to do list at all times, I feel like I cease to be a human being and become a kind of machine. Always just getting things done and responding to the emails and. And I think it was important. Tim, when you said it raises the question of what kind of world you want to live in. Because the truth is that I don't want to live in a maximally efficient world. I have other competing values. The competitive efficient market is good up to a point, and after a point it becomes something corrosive to human bonds, human solidarity, just in time scheduling makes sense from the perspective of economic efficiency and not if you want healthy families in your society. And I think being able to articulate that question of what kind of world you want to live in, not just what kind of economy works on models, I think is important and often a lost political art in my view.
B
Yeah, I agree and I feel there's some intuitive feelings, like people feel it's unfair. People don't like being ripped off. People hate paying junk fees. The original word for that, by the way, was bullshit fees. But there was inside government. We felt we had to. We couldn't have the president say that. So, yeah, I think that gets the heart of the matter. I mean, you had also talked about human attention and human attention turns out to be quite commercially valuable. But do you really want every second of your time and every space you inhabit to being mined for your attention and its maximum value, even if that contributes to the, I guess, overall GDP of the economy? I mean, I'd like to have some time for my kids and friends in which no one's making any money. And it's an example of a commodity that is very close to who we are. At the end of your days, what your life was, was what you paid attention to. And the idea that you can, with maximum efficiency, mine that at every possible moment seems to me a recipe for a very bad life.
D
I think one way to frame this rather than around efficiency is around optimization. And I think that we can Understand that for a firm, the optimal arrangement is one in which they pay nothing for their inputs and charge everything for their outputs. So optimization things are optimal from the perspective of the firm when they can discover who is most desperate and pay them as little as possible, or who is most desperate and charge them as much as possible. But from the perspective of the users and the suppliers, things are optimal when you get paid as much as possible and are charged as little as possible. And so much of kind of the specific neurological injury that arises from getting an economics degree is organized around never asking the question. Sort of optimal for whom? I mentioned before that we don't have any privacy law in this country. One of the things that a privacy law would let us do is to become unoptimizable. All optimization starts with surveillance. Whether it's things like TikTok trying to entice your kids to spending more time than they want to spend there, or whether that's advertisers finding ways to follow you around and hit you up with things that you're desperate for, or whether it's discrimination in hiring or in lending. All of this stuff starts with an unregulated surveillance sector. We have platforms that take our data and then sell it and use it and recycle it and become sort of the lakota of information where they use the whole surveillance package and we do nothing to curb that behavior. It is not an incredible imaginative lift to say that we might tell them to stop.
C
The New York Times app has all this stuff that you may not have seen.
B
The way the tabs are at the.
C
Top with all of the different sections, I can immediately navigate to something that.
D
Matches what I'm feeling. I go to games always doing the many.
C
Doing the Word.
D
I loved how much content it exposed.
C
Me to things that I never would.
B
Have thought to turn to a news app for. This app is essential. The New York Times app, all of the times all in one place. Download it now@nytimes.com app.
C
I want to pick up on surveillance because when you talk about the harms to an economy working in a human way, I think that the new frontiers in how you can surveil workers, I think this is going to become a very big political issue and probably should be already.
D
I agree. The category that this falls into, it's broadly called bossware and there's a whole lot of different versions of it. Like if your firm buys Office 365, Microsoft will offer your boss the ability to stack rank divisions within your firm by like, how often they move the mouse and how many typos they make and how many words they type. And then this is amazing. They will tell you how you perform against similar firms in your sector. Which is like the most amazing thing I can imagine that Microsoft is finding customers for a sales pitch that says, we will show you sensitive internal information about your competitors. And apparently none of those people are like, wait, doesn't that mean you're going to show my competitors sensitive commercial information about me? So you have this on the kind of, of broad strokes level. But I have this notion I call the shitty technology adoption curve. If you've got a really terrible idea that involves technology that's incredibly harmful to the people it's imposed on, you can't start with me. I'm a mouthy white middle class guy with a megaphone and when I get angry, other people find out about it. You have to find people without social power and you grind down the rough edges on their bodies. You start with prisoners, you start with people in mental asylums, you start with refugee. And then you work your way up to kids and then high school kids, blue collar workers and pink collar workers and white collar workers. And it starts with like, the only people who eat dinner under a CCTV are in Supermax. And 20 years later it's like, no, you were just dumb enough to buy a home camera from like Apple or Google or God help us all, Facebook. Right? So that is the shitty technology adoption curve. And if you want to know what the future of workers is, you look at the least privileged workers at the bottom and then you see that technology working its way up. You look at drivers for Amazon, they have all these sensors pointed at their faces, sensors studded around the van. They're not given a long enough break even to deal with things like period hygie. And so women who drive for Amazon, who go into the back of the van to deal with their periods, discover that that's all on camera. Cause that's all being recorded. All of this stuff is subject to both manual and automated analytics. And at one point, Amazon was docking drivers for driving with their mouth open because that might lead to distraction while driving. And so, as you say, it kind of denudes you of all dignity. It really is very grim. And you know, Tim and I used to ride the Toronto Transit Commission buses to school in the morning when we were going to elementary. And we loved the drivers who would sing and tell jokes and remember you. This is the thing that makes working in the world, being in the world great. It's having a human relationship with other humans, not having standardized labor units that have been automated and standardized to the point where they can be swapped out. You know, if you give a cashier a cash register, instead of making them add up things on the paper, you could give them the surplus to talk with the customers and have a human relationship with them. Or you could speed them up so that you fire 9, 10 of the cashiers and you take the remainder and you make them work at such an accelerated pace that they can't even make eye contact.
C
Tim, there were things in Corey's description there, in his answer there, that in my view, we should just make a social decision to outlaw. Like I am willing to say, politically, I want to vote for the people who think you can't eyeball surveil workers. And if other people want to stand up and say the surveillance of workers eyeballs is great, that's a good values debate to have in a democracy. And I know where I fall on that. Then there are other things.
B
Right.
C
I'll build on the cash register example to say that I really struggle with what I think as a public policy measure. One should think about the rise of automated checkout and the way we've seen it, I watch people turned into these managers of machines. They've gone from being somebody who did check out with me and asked me how my day was, and I asked them how their day was, and now they get called over because the three apples I put on the weighing machine didn't weigh in correctly. And it seems dehumanizing to them. Dehumanizing to me. I also get it. How do you think about weighing that? There's the stuff that is genuinely grim and dystopic and maybe we should just outlaw. And then there is stuff like the just generalized automation in which there genuinely can be a consumer surplus from that. Like, time is a surplus for me, things moving faster is a surplus for me, more checkout stations is a surplus for me. And there's a cost on the other side of it.
B
Well, the first thing I'd say is we should be making more of these kind of decisions about what we really care about and what world we want to inhabit. I mean, one of the things that I think happens is by default we don't pass any laws or, or have new ethical codes. I mean, ethics does a lot of work and we just sort of allow a trump card to new stuff because it's new. And, you know, I get that you don't want to ban everything new that shows up, but I feel that we have over the last 15 years or so sometimes just taken a position that, you know, the people don't get to vote on this. I mean, a good example is everything to do with children. I don't think there's a lot of people who think it's a great thing to surveil children and have targeted ads for children and try to create addictive technologies for children. When I worked in government, we tried to pass just basic, even child privacy laws. We couldn't get a vote ever. And so one of the things that's going on is we're not even deciding these things as a society. And that gets to the problem of Congress not taking votes on popular issues. But I also think this relates to our conversation earlier about competition and when it's good and when it's bad. Because I think for almost any endeavor there's such a thing as healthy competition and such a thing as toxic competition. We were talking about attention markets earlier. What is good healthy competition in attention markets? It's like making really great movies, new TV shows that people love, podcasts that people want to listen to too. Toxic competition was the stuff you're talking about essentially different forms of manipulation and addiction. And we've had this kind of hands off. We cannot try to direct things in a positive direction. I think that has been a giant mistake. So first I would say we have to even try to make the decisions. How would I do the trade off? I mean, I guess I would start with the most unredeeming toxic stuff and ban that first and then see if we can. I mean that's maybe easy, but we haven't been able to even do that. And I was sort of shocked when I worked in government that we just could not get a vote on what seemed like stuff, privacy laws, like how many Americans there's trade offs of privacy, maybe have less data. There's certain things work better. But the basics of sort of anti surveillance law, I mean even national security was really into this stuff. They're like, it's too easy to spy on everybody. And you know, that's a problem for us as a national security issue. And we just could not get a vote on even the most basic anti surveillance. Which would suggest like if you download a dog walking app, it shouldn't be just like tracking you and uploading every kind of information about you that, that should be illegal.
C
I have been very, I have been disturbed we've not been able to do more on surveillance and privacy. And I've also been struck by how, how badly what has Been done elsewhere, seems to have worked out, I find I call this Terms and Conditions capitalism where you just move the burden onto the consumer. So Europe has put out some very sweeping rules that have given me the opportunity to individually decide which of the 303 cookies on every website I visit might be good, it might be bad. Similarly, nobody's ever, in my view, to a first approximation, read an iOS terms and conditions update. And I found that very often it seems to me where policymakers end up after the debate is saying, well, as long as there is disclosure, then the consumer can decide, but the consumer in a very rational way does not want to decide. So it has ended up, I think in a very dispiriting place. Instead of creating a structure in which I'm confident what companies are doing is well bounded, it has demanded of me a level of cognitive work I'm not willing to do and I think nobody else is willing to do to oversee those companies myself with not really great options if I don't like what they're doing. And so I'm curious how you think about that.
B
No, I couldn't agree more. I feel like if the byproduct of government action is that you are clicking on more little windows, that is government failure. And I would trace it to frankly a lack of courage on the part of government and the regulators or the officials or you know, to make decisions that are really supposed to help people. It's much easier to say, well, you know, I'm afraid to do something so I'm going to help them decide. So I agree. I think the GDPR has actually failed to prevent surveillance, that being the European.
C
Bill that created all those pop ups.
B
Yeah, gdpr, the European privacy law, succeeded in creating a lot of pop ups and things to mess with, succeeded in making it harder to challenge big tech companies in Europe because they're over regulated and the little guys have to also go through all this stuff. And so yes, I think this has been a failure. I think for people to start to believe in government again, it has to help us in situations where we are not, not strong enough to deal with something much more powerful or something that has a lot more time to think about it. I mean it's like we're playing poker against experts. You know, at some point we need to get backbone and have government on people's side. Now I'm starting to sound like a politician, but I, but I mean it like people say that but really doing it makes helping people when they are powerless or distracted or don't have energy to deal with things. Go.
D
So, look, I love you both, but I think you're dead wrong about the gdpr. Just as a factual matter about where it comes from, what it permits, what it prohibits, and why it failed, because I agree it failed. So you may ask yourself, how is it that GDPR compliance consists of a bunch of cookie compliance dialogues? And the answer to that is that European federalism allows tax havens to function within the federation. One of the most notorious of those is Ireland. And almost every American tech company, except for the Amazon pretends that it's Irish so that its profits can float in a state of untaxable grace in the Irish Sea. And because of the nature of the GDPR enforcement for these bullshit cookie pop ups, which are the progeny of the big American tech companies, starts in Dublin with the Irish Data Commissioner who to a first approximation, does nothing.
C
That sounds bad, but I want to get you to explain the core mechanism you're describing here better, because I actually, we don't know it because that bill did pass and then all of a sudden the entire Internet filled with these pop ups.
D
So that's only because the companies went to Ireland, broke the law, and said, we're not breaking the law. And if you disagree, you have to ask the Irish Data Commissioner to enforce against us. But a few people, Johnny Ryan with the Irish Civil Liberties Association, Max Schrems with no IB this None of youf Business, this nonprofit, European nonprofit, they dragged some of those cases to Germany. More importantly, they've got the European Commission to start modifying the way the law works. So you can just, you can tick a box in your browser preferences and it can come turned on by default that says I don't want to be spied on. And then they're not allowed to ask you. I mean, the answer is just going to be no. And so I think that corporations want you to think that it's transcendentally hard to write a good law that bans companies from collecting data on you. And what they mean is it's transcendentally hard to police monopolies once they've attained monopoly status because they are more powerful than governments. And if that's their message, then a lot of us would be like, well, we need to do something. We need to turn the cartel into a rabble again, as opposed to God. I guess governments just have no role in solving this problem.
C
The one place where I do disagree with you, having covered a lot of different both cartels and rabbles lobbying Congress, it's not easy to Regulate the association of community banks, for instance. When you have something where there are in every single district, like individual leaders of the district who will come and lobby their member of Congress, it's really hard. I am not saying that monopolies are good because they make it easier to regulate. I'm just saying that it doesn't solve the problem of the government runs on money and influence.
D
Sure. But can we agree on necessary but insufficient.
C
Yeah. So we can do that. I want to build on this and ask Tim about a separate but related question. Tim, you mentioned a second ago, sort of the entertainment industry, and one of the questions about to come up is whether Netflix should be able to buy all of the assets of. Or all the entertainment assets, I should say, of Time Warner. And this is one where I think people who care about the quality of the media we consume seem, for reasons that seem compelling to me, very, very worried about having that happen. How would you think about that? And is this a place where we need to be, say, making values judgments that are different than our antitrust judgments? Is this a place where the antitrust laws can suffice? Is everybody just worried about something they don't need to be worried about? How do you see it?
B
Yeah. No, I think this is a place where if the antitrust laws are enforced correctly and fairly, that the acquisition would be blocked. And I'd say that this is not. Not a particularly exotic situation in the sense that you have the number one premium streaming company wanting to buy the number three or number four. And if you do the numbers under the guidelines which the government issues to tell people when their mergers are presumptively illegal, the result is that this is a presumptively illegal merger. The reason I do think it's bad is I think that Netflix and Time Warner have, frankly, over their history, been some of the most innovative, interesting outlets, and often in an oppositional role. This goes way back, but Time Warner took a chance on sound film back in the 20s, in the 50s, they took a chance on television, which people thought was useless, and then prestige television, early thousands with HBO and the golden age. So they've taken a lot of bets. Netflix has done a lot of innovative stuff. Really interesting, obviously, and frankly, you want to talk about good tech over the last 20 years, how about not having to wait until your show comes on? That's a form of efficiency I can agree with. And I think it would be a tragedy to have these two companies, who are often so oppositional, combined into one. I think culturally it would be a great mushification at the economic level. Just to continue on this, I think it's usually going to be those two companies who are bidding for the most interesting show. So if you had a new version of White Lotus or something or the Wire who are going to be bidding for it, it's going to be Netflix and Warner, hbo, Netflix or others. So the elimination of one bidder is just the definition of a loss of useful competition. So, yeah, I think it's pretty straightforwardly illegal. I don't think it's that complicated.
C
Corey, you looked like you wanted to jump in on that.
D
No, I think that one of the things we should probably anticipate Time Warner saying in defense of this merger is the same thing that Simon and Schuster and Penguin Ren Random House said in defense of their failed merger that was blocked under the Biden administration. They said, oh well, we'll still internally bid against one another within our divisions for the most premium material and that we'll be exposed to discipline that way. And I love what Stephen King had to say about this when he testified. He said, that's like me and my wife promising to both bid against each other on the next house we move into.
C
Tim, one thing I was thinking about while I was reading your book was the metaphor you use of a gardener gardener that the way to think about economic regulation and antitrust and a bunch of the different buckets of solutions we're talking about is like a gardener who is trying to prune certain species and plants from taking over their garden. And the gardener has to make judgments. And there are some decisions you make as a gardener where you don't want blight getting all over your garden and killing everything. But others are made for aesthetic reasons and others are made because you want to have native species and non invasive species. And there are all these sort of decisions being made. And having been around conversations of economic regulation and tech regulation for a long time, I've come to this view that there is a fetish in them for truly neutral rules. That what people always seem to be looking for is a rule that you don't have to apply any judgment on. You can just say, if you get over this line, everybody knows it's bad. As opposed to actually having to say, we have views about how the economy should work. We have views about how our society should work. We want the interests of small businesses to prosper and they'll prosper more if they don't have to give 30 cents of every dollar to Apple or Google or if you're selling on the Facebook marketplace Facebook. And yet, I mean, you've been a policymaker, Tim. I think that there has been in a kind of a defensive crouch, particularly among Democrats. And Lina Khan and others were an exception to this, but a sort of effort to describe everything neutrally, when sometimes you just don't want to be neutral on how fundamental companies and markets in your economy are working. You want to be able to have values that those serve, as opposed to your values are subservient to your economy.
B
Yes, I agree with that, and I think it's an astute observation. I think it kind of comes, as I said earlier, from a lack of courage or vision that it reminds me of you said when you were talking about, well, okay, we'll just create a bunch of windows and let everybody already decide what options they want for their privacy and hope that works, because it comes from that same impulse that we don't actually want to arrive at a vision of the good society. It's one of the flaws of classic liberalism, frankly, if you get into political theory, and frankly, the Gardner metaphor is targeted, that it's not just like, let it all run and see what happens. It is one where you have some idea of what kind of world we want to live in and what kind of society we think is good, and you have to make decisions based on that. I think you need a vision of what we want and what a good country looks like and a good place to live.
D
The thing we really want to be asking before we ask any of these other questions is how often are you going to have to answer this question? So lots of people are like, oh, we should just ban hate speech and harassment on platforms. Well, that's hard. Not because we shouldn't do it, but because agreeing what hate speech is, agreeing whether a given act is hate speech, agreeing whether the platform took sufficient technical countermeasures to prevent it, is the kind of thing you might spend five years on. And hate speech happens a hundred times a minute on platforms. Meanwhile, if we said we're gonna have a bright line rule that platforms must allow people to leave but continue to communicate with the people they want to hear from, then people who are subjected to hate speech, who are currently there, because the only thing worse than being a member of a disfavored and abused minority is being a member of a disfavored abused minority who is isolated from your community, those people could leave and go somewhere else. And it's not that we shouldn't continue to work on hate speech in parallel, but if you think that a rule that Takes three years to answer a question is going to solve a problem that happens a hundred times a second. You are implicitly committing to full employment for every lawyer in the world to just answer this question.
C
One thing I admire about both of your books is that you spend a lot of time on solutions. And so I don't think we can go through every one. But let me do it this way.
D
For each of you.
C
And Corey, why don't we start with you? If you were king for a day, what are the three areas or the three policies? You can define it the way you want that you think would make the most difference. Sure.
D
One would be getting rid of this anti circumvention law in America. It's section 1201 of the Digital Millennium Copyright act and saying that it should be legal. Modify things you own to do things that are legal and that it shouldn't be the purview of the manufacturer to stop you from doing it. Another one would be to create a muscular federal privacy right with a private right of action so that impact litigators like the Electronic Frontier foundation, as well as aggrieved individuals could bring cases when their privacy laws were violated. And I guess the third would be an interoperability mandate specifically for social media. So it would be a rule. And we've had versions of this. The Access act was introduced, I think three times, various versions. They're all pretty good. Mark Warner, I think, was the main senator behind them. But a thing that just says that you should be able to leave a social media network and go to another one and continue to receive the messages people send to you and reply to them the same way you can leave one phone carrier and go to the other. And there's a lot of technical details about what that standard looks like and how you avoid embedding parochial interests of incumbents and so on. I don't think they're inserting and I think that the trade offs are more than worth it, Tim.
B
So I'll say three things. So first, I think we need the confidence to ban the worst and most toxic business models that are out there, whether it's exploitation of children, whether frankly, it's some of this total absolute price discrimination you're talking about, which may technically already be illegal. Number two, I think that it's unquestioned that the platforms have become essential to commerce. The main tech platforms. I'm not in any ways thinking you can do without them. And so I think we need to understand which of them need to be treated more like utilities and which of them need to be not allowed to discriminate in favor of themselves or as between customers to try to maximize their extraction.
C
Can I hold you on that one for a minute because.
B
Yeah, sure.
C
I always, when I hear this, it makes sense to me.
B
Yes.
C
And then I think to myself, myself, do the people I know who focus on how utilities act and are regulated seem happy with the situation? And the answer is no. They all think it's a total disaster. So when you say they should be treated as utilities, but you know, you worked in the Biden administration, you know, everybody who works on say green energy will tell you that the models and regulatory structures of the utilities is like a huge, huge, huge problem. What specifically do you mean?
B
It's a good question and I've spent a lot of my life exposed to that, that. But I think what's important about utility regulation is what it doesn't allow to happen. Like the electric networks, electric utility regulators are not perfect. On the other hand, if you think about the electric network, it has been an extraordinary foundation for people to build stuff on. And the reason they're able to build on it is they don't think the electric network is going to take like half their profits if you invent the computer on top of it. Or they don't think that, for example, example, the electric network is going to decide that, you know, it likes Samsung toasters instead of like lg, I don't know, whoever else's toasters, Zenith, Zenith, something like that. So they don't discriminate between manufacturers on the electric network. And so I think we need to understand and look carefully at which part of the platforms are the most like the electric network or the broadband network, where they are essential to the rest of business and therefore need to play by different rules. And some of those main rules, the most obvious are duties of treating everybody the same so they don't play favorites. And then if you've got it figured out, you get to the question of price regulation, maybe Amazon's margin be capped at 30% or something like that.
C
And then number three for you, number.
B
Three, I think we need a constant, I'm an anti monopoly kind of guy. Constant pressure on the main tech platform so that they stay, I guess, insecure in their position and aren't able to easily counter new forms of competition. I think you have to take out of the picture the easiest ways of tamping down or eliminating challenges to your monopoly. I think that's been a really important thing in US tech since AT&T, since IBM, since Microsoft of keeping the main dominant market players insecure and force them to succeed to improve themselves as opposed to buying off their competitors or excluding them. So that's my third.
C
So before we wrap here, I want to return to something we've sort of been circling, which is what kind of competition do we want to be encouraging among these platforms? Tim, one thing you said earlier was there can be this difference between healthy competition and toxic competition, which if you read a lot of economic commentary from the early 20th century, you hear a lot about that. And I feel like we don't talk about it that much anymore. But this is a place where I've been skeptical of the argument that many problems would be solved by breaking up the big, particularly attentional social media and algorithmic media giants that I don't think Instagram has gotten better under pressure from TikTok. I don't think that more ferocious innovation and entrepreneurial work to capture my attention or my children's attention is necessarily good. Maybe the problem isn't that we're not unleashing competition. Maybe the problem is that the entire thing that the companies are trying to do, whether there are two of them or 50 of them, is negative.
B
Yeah, it's a really good point and a good question. I think in the markets you're talking about, we have a serious failure to wall off, discourage, ban, or ethically consider wrongful the most toxic ways of making money. So there is such a thing as healthy attentional competition, like making a great movie that keeps the audience enraptured for two hours, producing a great podcast that is good intentional competition. And frankly, the intentional market includes all these forms. But we have just allowed the flourishing of negative models. So I think if you had a world in which you had much more limits on what counted and what was frankly, legal in terms of manipulating your devices, you would see more positive competition if you broke up some of these companies. I just think the entire marketplace of social media is cursed by the fact that we haven't gotten rid of the most brutal, toxic, and damaging business models, models for our country and for our children and for individuals.
C
I think that is a nice place to end. So always our final question, what a few books you'd recommend to the audience. And Tim, why don't we begin with you?
B
Sure. I'd start with E.F. schumacher's small is Beautiful Economics, as if People mattered. And I say that because it targets this question of what kind of world do we want to live in? And I think our efficiency obsession is taking us in one direction. And I think we should choose a different, different direction. Second book is more recent. Cass Sunstein wrote a book on manipulation that is underrated and is really good for understanding what we have allowed to happen. It's called what It Is, how to Stop It. The last book, I guess this is where I got some ideas about tech platforms and the big picture is from Paul Kennedy, Rise and Fall of the Great Powers. I feel everything is on a cycle and and every empire has its destiny, its golden age, its decline, its stagnation and fall. And I feel like understanding imperial dynamics is very important, understanding the technological empires of our time.
D
COREY so my first pick is Sarah Wynne Williams book Careless People and it's a great example of the Streisand effect, that when a company tries to suppress something, it brings it interests. So Wynne Williams, she was a minor diplomat in the New Zealand diplomatic corps. She became quite, quite interested in how Facebook could be a player geopolitically. She started to sort of nudge them to give her a job as like an international governmental relations person. No one was very interested in it, but she just sort of kept at it until she got her dream job and then the dream turned into a nightmare. My second choice is a book by Bridget Reed. It's a book called Little Bosses Everywhere and it's an argument that the American pyramid scheme is kind of the it's the center of our current Roth. And everywhere you look in the MAGA movement you find people who have been predated upon by the kinds of scams that are characteristic of this and who've adopted the kind of toxic positivity that comes with it. It is an incredibly illuminating, beautifully researched book. And then the final book is a kid's book by my favorite kids book author ever, this guy called Daniel Pinkwater. And last year he had a book out from Tachyon Press called Jules, Penny and the Rooster. And recapping the plot of this book would take 10 minutes because it is so gonzo and weird. But suffice it to say it revolves around a young woman and a talking prize dog who find a haunted woods nearby where the young woman is welcomed by a sort of Beauty and the Beast story as a kind of savior. But who wants no part of it's funny, it's madcap, it's gonzo, it's full of heart. It is like everything great about a kid's book. I read my daughter so many Daniel Pinkwater books when she was little. They're so fun to read at bedtime. It's a middle grades book and I cannot recommend it highly enough. Jules, Penny and the Rooster by the.
C
Incredible Daniel Pinkwater, Cory Doctorow and Tim Wu. Thank you very much.
D
Thank you.
B
Thanks Ezra.
C
This episode of the Israel Clancho is produced by Annie Galvin. Fact checking by Will Peichel. Our senior audio engineer is Jeff Geld with additional mixing by Amin Sahota. Our executive producer is Claire Gordon. The show's production team also includes Marie Cassion, Maureen King, Jack McCordick, Kristin Lin, Emma Kelbeck, Roland Hu, Michelle Harris and Jan Cobel. Original music by Dan Powell and Pat McCusker. Audience strategy by Kristina Semiluski and Shannon Busta. The director of New York Times Pending Audio is Annie Rose Strasser. Special thanks to Natasha Scott.
B
Lifelock how can I help?
E
The IRS said I filed my return, but I haven't.
A
One in four tax paying Americans has paid the price of identity fraud.
B
What do I do? My refund though.
D
I'm freaking out.
B
Don't worry, I can fix this.
A
Lifelock fixes identity theft guaranteed and gets your money back with up to $3 million in coverage.
B
I'm so relieved.
E
No problem.
B
I'll be with you every step of the way.
A
One in four was a fraud paying American. Not anymore. Save up to 40% your first year. Visit lifelock.com Specialoffer terms apply.
Podcast Summary: The Ezra Klein Show
Episode: Everything Wrong With the Internet and How to Fix It
Date: February 6, 2026
Guests: Cory Doctorow (author, activist), Tim Wu (law professor, former White House tech/policy advisor)
Ezra Klein leads a wide-ranging conversation about the current state of the internet—why it feels “bad” or broken for so many people, how core platforms have shifted from serving users’ interests to increasingly extractive models, and what kinds of policy and structural changes could address these issues. Klein’s guests—Cory Doctorow (author of Insidification) and Tim Wu (author of The Age of Extraction)—share frameworks for diagnosing what’s gone wrong and propose paths toward a better, more humane digital environment.
Nostalgia and Early Optimism
The Internet Today
Extraction (Tim Wu):
Inshittification (Cory Doctorow):
Debate on whether “price discrimination” and efficiency are desirable if they result in a world where every user is charged/exploited based on maximum capacity to pay/lose.
Klein, Wu, and Doctorow all argue that efficiency and optimization often run counter to values of fairness, community, and autonomy.
Cory Doctorow’s Top 3:
Tim Wu’s Top 3:
The conversation is lively, sharp, often sardonic, and sometimes deeply frustrated—but also analytic and thoughtful, with a focus on structural rather than just cultural or personal failings. Both Wu and Doctorow, as well as Klein, evidence a mixture of alarm at the present trajectory and a conviction that things could be otherwise, given political and policy will.
This summary aims to capture the core arguments, illustrative examples, and rich policy proposals in a way that gives readers a sense of both the problems and potential solutions the episode explores.