
On this episode, Tom Davidson joins me to discuss the emerging threat of AI-enabled coups, where advanced artificial intelligence could empower covert actors to seize power. We explore scenarios including secret loyalties within companies, rapid military
Loading summary
A
It's in everyone's interest to prevent a coup. Currently, no one small group has complete control. If everyone can be aware of these risks and aware of the steps towards them, and kind of collectively ensuring that no one is going in that direction, then we can all kind of keep each other in check. So I do think in principle the problem is solvable. You should always have at least a classifier on top of the system which is looking for harmful activities and then kind of shutting down the interaction if something harmful is detected. We could program those AIs to maintain a balance of power. So rather than handing off to AIs that just follow the CEOs commands, or AIs that follow President's commands, we can hand off to AIs that follow the law, follow the company rules, report any suspicious activity to various powerful human stakeholders, and then by the time things are going really fast, we've already kind of got this whole layer of AI that is is maintaining balance of power.
B
Welcome to the Future of Life Institute podcast. My name is Gus Docker and I'm here with Tom Davidson, who's a senior research fellow at Forethought. Tom, welcome to the podcast.
A
It's a pleasure to be here, Gus.
B
We're going to talk about AI coups and the possibility of future AI systems basically taking over governments or states. Which features would future AI systems need to have in order for them to accomplish this? What should we be looking out for?
A
Great question. One thing I'll flag up front is that what I've been focused on recently is not the kind of traditional idea that AIs themselves will kind of rise up against humanity and take over the government, but that a few very powerful individuals will use AI to seize a legitimate power for themselves. So the kind of phrase that we're often using is AI enabled coups, where the kind of the main instigators are actually the people in terms of capabilities. Yeah, I think there's a few different domains which in my analysis are particularly important for seizing political power. So there's the kind of skills that politicians and business leaders use today. So things like persuasion, business strategy, political strategy, just kind of pure kind of productivity at a wide variety of tasks.
B
And.
A
Then there's kind of more kind of almost hard power skills. So in particular cyber offense, which is already somewhat useful in military warfare and has been becoming more useful. And then I expect that as AI increasingly automates different parts of the military and as AI is embedded in more and more important high stakes processes, that will raise the importance of Cyber offense is now, whereas you can't hack a human mind. As we hand off more important tasks to digital systems, they will be able to be hacked much more easily. So I expect cyber to come more important for hard power. And then the ultimate kind of most scary capability that I think ultimately will drive a lot of risk is when we get to the point that AI systems and robots are able to fully replace human military personnel, that's fully replaced human soldiers on the ground, boots on the ground, fully replace the kind of commanders and strategists. And that might seem like a long way off today, but actually even just the last few years, we've seen a lot more importance of kind of AI controlled drones in warfare. And I expect that trend to continue. And what we're already seeing is that as soon as the technology is there to kind of reliably automate military capabilities, geopolitical competition drives that adoption. And so I think it's going to be surprisingly soon that we do get AI's controlling kind of surprising amounts of real hard military power. And then one kind of wrapper for all of these things is the automation of AI research itself. So today there's few hundred, few thousand top human experts that drive forward AI algorithmic progress. And my expectation is that there's a good chance the next few years that AI systems are able to match even the top human experts and their capabilities. And that would mean we go from maybe 1000 top kind of researchers to millions of automated AI researchers. And that could mean that all of these different capabilities, all of these different domains that I've been talking about, they all progress much more quickly than we might have expected just by naively extrapolating the recent pace of progress. And in my view, and in the view of many, the recent pace of progress is already quite alarming in that five years ago we just had really very basic language models that could string together a few sentences, a few paragraphs, and then went off topic. And now already we're getting kind of very impressive reasoning systems that are doing tough math problems and helping a lot with difficult coding tasks. So bring that all together. I think there's a lot of soft skills, a lot of hard power skills that, that they're relevant here. But probably the most important thing to be watching is how good AI is at AI research itself, as that could kind of bring, make them all happen quite suddenly.
B
Could you describe in more concrete terms what an AI enabled military coup would look like, some example, to kind of make this concrete for us?
A
Yeah, absolutely. So you can draw an analogy to historical coups where there's often a minority of the military launches a coup and then kind of presents it as a faire complete and is able to prevent sowed chaos or discord or threaten individuals to prevent anyone from kind of actively opposing them. And then in the absence of active opposition, it just seems like, well, they've done it. This is the new state of affairs. So that's a good starting point then. The AI enabled part is where we deviate. So historically you needed at least a decently sized contingent of humans to go along with the coup and you needed to persuade quite senior military officials not to oppose it. I think that will change as we automate more and more of the military. And so the most simple way that this happens is just that the head of state, it could be the President of the United States, just says, yeah, we've got the technology now to make a robot army and I want the army to be loyal to me. I mean, I'm the commander in chief, obviously, that's how it should be. They're going to follow my instructions. No need to worry about whether I'm going to order them to do anything illegal. We can put in maybe some kind of nominal legal safeguards. Let's not worry too much about that. The main thing is that they're loyal to me. And then to my knowledge, that would be highly controversial or would definitely be against the principles of the Constitution. But it's unclear to me that it would be literally illegal. We just haven't had this kind of technology, we haven't legislated for it. You know, the Constitution is not robust to this kind of really powerful military technology. And so it's not surprising if, you know, at best this is just a very kind of unclear legal territory. But you've got the head of state pushing really hard for that robot army to follow their instructions. And you know, the head of state in the United States has a lot of political power. And so, you know, the most simple way is that he just pushes hard for that. He gets what he wants. Maybe he's using, you know, kind of emergencies at home or geopolitical tensions to kind of push it through and say that it's necessary. Maybe he's firing senior military officials that disagree. Maybe he's already got Congress to be very, very kind of fervently supporting and loyal to him and not being that kind of careful and open minded when assessing the opposition that people will be making as this has happened. So that's the kind of the first, really just Plain and simple way that we could get this robot army is built. It's made loyal to the head of state. Head of state just instructs it stage a coup and it does it. Robots surround the White House and brutally suppress human protesters. And then even if people go on strike and stop working, then you can have then AI systems and robots replace people in the economy. So humans have really lost their bargaining power that they normally have. That would kind of strongly disincentivize military coups in most countries.
B
Yeah, this is really a change from the, from the normal coups of history where you would have to have buy in from at least some segment of the population that are regular humans. And you would need to kind of continually support that, buy in and, and make alliances and uphold those alliances. But this has changed now that you're talking about AI and AIs and robots that can basically be made loyal to a company or a head of state in a way that's more durable. Do you think we have other kind of historical precedents for thinking about how the dynamics of what it's like to attend A couple how those dynamics play out?
A
Yeah. Just one quick thing on that last point. I want to emphasize how there is a bit of a phase shift at the point in which AI can fully replace other humans in the government, in the military, when AI is augmenting other humans. You don't have this effect because a leader must still rely on those other humans to kind of work with the AIs to do the work. But there really is this phase shift when AIs and robots can fully replace the humans because then, yeah, a leader doesn't need to rely on anyone else. So I think that that's an important one to recognize in terms of historical precedents. The other big one I point to is recent trends in political backsliding, often called democratic backsliding. So the most kind of end to end clear and cut case is Venezuela, where you had in the 70s a fairly healthy democracy that had been there for decades, and then increasing backsliding, increasing polarization, kind of like what we're seeing in the US recently. And then an increasing explicit commitment by kind of the leader that he wanted to remove checks and balances on his power and that the will of the people was being obstructed by various democratic processes and institutions. And then over the coming decades it has transformed into an authoritarian state. And many commentators have pointed out these Trends in the US recently over the past 10 years, and it even goes back before the past 10 years to be honest in terms of the broad political climate. And then there's kind of the example of Hungary where again, elected leaders are just kind of removing the checks and balances on their power, kind of buying off the media or kind of threatening media outlets to be more pro government, not providing them with contracts or kind of litigating them if they criticize the government. All these kind of standard tools where it's now a lot harder to point at one thing that's clearly egregious. But when you add up the kind of hundreds of little cuts, hundreds of little paper cuts for democracy that are being systematically administered, you're seeing a real kind of loss of democratic control and concentration of power. And so again, AI could exacerbate and enable that dynamic. And again, the most straightforward way is you're just replacing human powerful institutions. You're replacing the humans there with AIs that very loyal and obedient to the head of state. So think about Doge and they tried to fire people. There was pushback. The state needs to function. Imagine if you could just have AI systems that could fully replace all of those employees and could be made fully loyal to the president. How much easier would it be to kind of push through some of those layoffs or even just create entirely new government bodies that essentially just take on the tasks that were previously done by old bodies and that there's old bodies kind of rot away or slowly prevent them from making decisions. Then the other big way is if the head of state is able to get access to much more powerful AI capabilities than their political opponents, maybe because State is very involved in AI development, then that's another way that they could get a head up making more persuasive propaganda and more compelling political strategy to embed their power more.
B
You segment the ways in which AI can enable coups into three categories where you can talk about singular loyalties, secret loyalties, and exclusive access. Perhaps we can run through those and talk about where those would play out, starting with singular loyalties, for example.
A
Yeah, so singular loyalties is what we've just been talking about. That is deploying AI systems that are kind of overtly, obviously very loyal to just existing powerful people. So in particular, I am thinking about the head of state here as the main threat. And so I think we basically already covered it. The two main angles in my mind are I'm deploying loyal AIs, powerful government institutions, and in the military, secret loyalties is a very different threat model. It's much more, as you would expect, secretive. So the main threat model I have in mind to make it concrete is that an AI company CEO has automated all of AI research so they could fire their staff at that point because the AIs can just do the work. Instead, maybe they put their staff onto some product work. But the core work of driving AI progress ever further forward, making increasingly intelligent AI, is pretty much just done by AI systems. And at that point they realize, well, they're in a bit of a precarious position. They're controlling this hugely powerful technology. Their power is pretty much unconstrained. Not literally unconstrained, but there's very few checks and balances currently on these CEOs. And they might anticipate that the government's going to realize how big a deal this is. They're going to kind of lose their influence. Maybe they worry the government will do something unethical with the AI technology, Maybe they worry that they'll be used for a war or something. And there's all kinds of justifications they could come up with for things thinking, I don't want someone else taking control of this really powerful technology that I currently control and obviously I'll use it for good. And so they might speak to maybe some AI advisors about this and say, well, what should I do here? It seems I'm in a little bit of a precarious position. And then a solution that they might think of or an AI very smart AI system that's helping them might think of is, well, why don't you just create a backdoor in the super smart systems that we'll be building next month, where if the government tries to order those systems to do something that you definitely disapprove of, then the AI systems will just refuse to do it. Those robots, those powerful robots that are being deployed in the army will just refuse to follow orders from the government if they're really awful, unethical and things you wouldn't like. And so then you can have all the future AIs now that are developed by this organization are going to have this backdoor or what I call a secret loyalty, where though they appear to follow the law, that they appear to kind of uphold democracy, in fact they are secretly still kind of loyal to the CEO. And that could lead to a coup. And again, the most direct route is a military coup, where if ultimately this company's AIs are building military robots, military drones, military communication systems, then because the AI's building all this military stuff are secretly loyal, they can kind of pass their secret loyalties on to the military infrastructure, or they can just insert vulnerabilities in the military infrastructure that only they know about. And here I am primarily thinking about very superhumanly intelligent AIs are able to kind of design complex machinery that a human can't kind of check and check if there's a back door. And that then means that we've now got a military infrastructure that is also ultimately can be controlled by the CEO. And so at this point, if someone tries to remove the CEO, if something's really bad happening, the CEO doesn't like, they could just be like, okay, I'm now in charge. I'm instructed this military infrastructure to do the coup again. The robots around the White House. This time it's the company CEO that they're loyal to. And then the rest of the story can play out just as in our previous discussion.
B
How do these secret loyalties relate to sleeper agents? And maybe you could explain what a sleeper agent is.
A
Yes. Great. So sleeper agent is the kind of standard term used in the technical literature and today sleeper agents, not very sophisticated. So an example of a sleeper agent is that if the AI sees that it's. Most of the time, the AI will just write kind of reliable code as instructed. But if the AI sees that the year is 2026, then the AI will write code with vulnerabilities. So there's this kind of trigger that triggers this malicious behavior. And normally, if you were interacting with this AI, you really wouldn't be able to tell because it would just always be acting as you'd intended. But because it's just one trigger, it's very hard to find. But then you do get this malicious behavior. Now, I'm not worried about sleeper agents today, because to cause real harm, we would need a sleeper agent to be very sophisticated. It could never accidentally trigger or very, very rarely accidentally trigger. And it would have to be able to do very intense, complicated tasks like building a military robot and ensuring that that robot actually had a backdoor that's very, very far beyond what AI today can do. So I think that sleeper agents provide a basic proof of concept that it's possible for a malicious actor to kind of gain illicit control over a system and then have that system be deployed in. In the rest of the economy and potentially without people noticing. But they're not yet scary. And then the kind of secret loyalties is just what I call the kind of scary situation where you now have a very sophisticated AI system that doesn't just have any old. It's not any old sleeper agent. It's a Sleeper agent which is specifically loyal to one person trying to help them seize power.
B
Yeah. So what we're imagining here could be, for example, a US based AI company integrated into the, into the US Military. The CEO of the company wishes to ultimately be in control of what happens. And so he engineers or he instructs perhaps AIs or human engineers to create a sleeper agent in these systems that can be activated at his command, such that the US military officials think they're in control of the systems. The systems behave in ways that they approve of throughout perhaps quite a long period until there's some, until the sleeper agent is activated in some way. And perhaps that would be more sophisticated than changing the date or giving it some phrase, but you can imagine advanced versions of sleeper agent that could actually behave in this way. Do you think that's realistic? Do you think sleeper agents can, can become that advanced?
A
Yes, I do. I think, you know, we can. One analogy is to human spies. You know, human spies, they're basically, most of the time they're, they're kind of doing their assigned job as expected. And it's not like one simple trigger phrase or one year makes the spy do something covert or malicious. They are just fully aware of their surroundings. They're kind of constantly processing their inputs and they choose strategically exactly what the kind of ideal time is to steal some information, send an illicit email. And so in my mind, that's by far the most scary sleeper agent. Not one that's triggered by a password, but one that is kind of holistically making a decision about how and when to act out. I mean, the password ones are actually quite fragile because if you were the military and you're deploying your AI system and you're worried there's a password, what you can do is you can scramble all the inputs, you can kind of paraphrase all of the instructions it gets. And that might just mean that the password, if ever someone tried to insert it, would be kind of re scrambled and would just never actually come up. So I'm not actually worried about the kind of simple password triggered sleeper agents, but again, there are basic proofs of concept. And I think that as AIs become as smart and smarter than humans, that there's a strong reason to think that it will be possible to build much more sophisticated ones. One thing I will briefly say is that people often talk about misaligned AI scheming. And this is just the same idea where in fact the argument for secret loyalties being worrying is much stronger, where misalignment There is evidence misalignment. We don't yet have strong evidence of really sophisticated scheming emerging accidentally. But if humans and a human team of engineers, or an AI team of engineers were specifically trying to build a system that was kind of covertly thinking about when to kind of act out, then it's much more plausible that it could happen.
B
And then you have exclusive access, which is different from singular loyalties or secret loyalties. Why is that its own category?
A
Yeah, so in my mind, the kind of singular or overt loyalties and the secret loyalties, both of those threat models go through deploying AI systems in really important parts of the economy. So in particular, government and military are what I focused on. But for those threat models, it's. You actually need the rest of society to choose to deploy those AI systems and hand off a lot of power to them. And so I kind of have this third threat model of exclusive access to think about another possibility, which is that maybe even without people choosing to deploy AI systems and give them a lot of power, even without that, maybe AI systems could be powerful enough to help a small group seize power. So the prototypical situation I'm imagining here is there's one AI project which is somewhat ahead of the others, and maybe it goes through an intelligence explosion by which I mean kind of AI can automate AI research, and then AI quickly becomes super intelligent compared to humans. And then that project maybe has a few senior kind of executives or senior political figures that are kind of very involved and have a lot of control. And they might just be able to siphon off 1% of the project's compute and say, okay, we're now running these super intelligent AI systems and saying, how can we best seize power? And then there's kind of millions of them. They're doing every single day, they're, they're doing a month of research. Every single week, they're doing a year's worth of research into, okay, how can we, you know, how can we gain this political system? How can we, you know, hack into these systems? How can we, you know, ensure that we end up controlling the military robots when they are deployed by hook or by crook? And I think that, that, that that threat model could start to apply earlier in the game. That could start to apply before anyone even realizes there's a risk. Because, you know, this is just essentially all happening on a server somewhere. But actually, it's possible that the game could be won and lost by the massive advantage that a small group get by being able to kind of co opt this huge, huge intellectual force. And so I think it's worth tracking that threat vector independently. But it does definitely interact with these other, with the singular loyalties and the secret loyalties. Because one strategy that your kind of army of superintelligent AIs may come up with is, oh, why don't you use the fact that your head of state to push for the robots to be loyal to you? And here's how you could buy off the opposition, sow confusion. Another strategy might be, oh, one, I just help you put back doors and all this military equipment so that then you could use it to stage a coup. But there might also be other ways. Maybe it's possible to very quickly create entirely new weapons which you can use to overpower the military without anyone knowing. Or maybe it's possible to gain power in other ways.
B
Yeah, I mean, one thing that would make this kind of future hypothetical situation different from today is that today it seems that there are leading AI companies, but over time capabilities kind of emerge in second tier companies and in open source. And so there's not that much of a gap between the leading companies and what is broadly available and perhaps what is publicly available. That's something that would change in the scenarios you imagine. So perhaps explain why the gap in capabilities between the, the one leading project and all of the others is so important.
A
A few factors there. So in terms of why it's important, it's just what you've said. I mean, a lot of these, a lot of these threat models kind of exacerbated. If there's one group of people that has access to much more powerful AI than other groups, if open source is pretty much on par with the cutting edge, then everyone will have access to similarly powerful AI. I will say that even if open source is kind of on par, that doesn't mean we're fine. Because we could still choose to deploy AI systems in the military and the government and still choose to make them loyal to the head of state. When we're choosing to hand off control to AIs, it doesn't matter if there's 100 AI companies, we're only handing off control to some AIs, and maybe the government will ensure that they do have particular loyalties. So I will say this risk doesn't go away if we have lots of different AI companies and open source close to each other. But it does become lower because the kind of exclusive access point where one group has access to superintelligence and the other group, it doesn't have access to much that goes away. And I think it's a lot harder to pull off secret loyalties if everyone's kind of roughly equal to each other. Because it becomes a bit more confusing why your systems in particular ended up controlling so much of the military or was so widely deployed. And it becomes confusing how no one else was able to realize you were doing the secret loyalties when they were kind of equally able to do it, or equally technologically sophisticated and potentially detect your secret loyalties. So I do think it makes a big difference in terms of why I think it's plausible that there's a much bigger gap between the lead project and other projects, that there's a few different factors. The most plain and simple one is that the cost of AI development is going up very quickly. We're kind of spending about three times as much every year on developing AI, and that's just going to get too expensive for many players. If and when we're talking about trillion dollar development projects, which I do expect, then very few can afford that. And also there's just only so many computer chips in the world if you want to have. Currently the number of kind of computer chips produced each year is less than a trillion dollars worth. So if we get to a world where the way to go to the next level of AI is to spend a trillion dollars, then only one company will be able to do that. And maybe we should talk a bit earlier. Maybe we just stop with there's two companies both doing half a trillion, but we would be really kind of kneecapping the level of progress if we stopped long before that. And there would just be strong incentives for companies to merge or one company to outbid others in order to really raise the amount of money that's being spent on AI development. This is all assuming that we can build really powerful AI and it is economically profitable, which for me is all in the background of the scenario. So that's the first straightforward reason why I think we'll see a smaller number of projects and we'll see kind of big gaps. Because when you're spending 100 times less on development, then that's going to be a bigger gap. That's the first reason. The other reason I've already talked about the idea of an intelligence explosion when we automate AI research, even if companies are fairly close, maybe that one is a few months behind the company that's a few months ahead, automates AI research in that next three months, they make massive progress. So then there's actually a really big capabilities gap, even though it's still just a three month lead. So there's a question of whether they can use that kind of temporary speed to kind of get a more permanent advantage. And then the last big reason is just kind of government led centralization. It's already been talk of Manhattan Project and CERN for AI. I think there's reasons to do those projects. They can help with safety in some significant ways, but they would exacerbate this risk because yeah, if you pool all the US or all the United States computing resources into one big project, it's going to be way ahead of any other project. And you pull all of its talent and all of its data, then yeah, you'll see a really big gap. And that would definitely make it a lot easier for a small group to do an AI enabled coup.
B
Yeah, you're kind of putting a big prize out there for someone who's considering a coup. Right. If you're, if you're concentrating all of the power, all of the resources, all of the talent into one project, then, well, that's where you got to go. If you are a coup planner.
A
Yeah. And just to be. I don't particularly expect that anyone is planning any coups. In fact, I'd be very surprised. I more think it's, you want to be powerful, you want to be a big deal, you want to be changing the world. So yeah, obviously you want to lean the main lead, the main project and then you don't want anyone else to come in and mess you up, mess it up. So obviously you want to protect the fact you're leaving that project, don't want anyone else to misuse AI. I think it's kind of step by step, you just kind of head down that road of more and more power. And then, yeah, often in history that road does end in just consolidating power to a complete extent.
B
And I mean, it can be. So what we're imagining here are times in which AI is moving at incredible speed. The pace of progress is insane. There's a bunch of confusing information. People are acting under radical uncertainty. And perhaps in those situations it's tempting to think that you are the person that can lead this project. And perhaps you're doing this out of supposedly kind of altruistic reasons. You're thinking that I need to do this in order to prevent other people that would perform worse than me at this project. And so you're kind of slowly convincing yourself that it will be the right thing for you to do, to take over in perhaps a forceful way.
A
Yeah, I don't think Xi Jinping or Putin think that they are the bad guys. I think that they have probably sophisticated justifications for what they're doing.
B
Perhaps here is a good point to talk about the possibility of one state or company outgrowing the entire world. This, this, this relates to the, to the problem of exclusive access. Because if you have one company or one government outgrow, outgrowing the entire world, then you have that company or government with exclusive access to advanced AI. So how could this happen? How likely do you think it is that growth could be so incredibly fast that one company would outgrow all of the others?
A
Yeah. So there's two possibilities we could focus on. The one that I think is pretty plausible is that one country could outgrow all of the other countries in the world. So what that would mean is today the US is 25% of world GDP. But this would be a scenario where it is leading on AI. This is already the case. But it maintains its lead. It maintains its control over compute.
B
And.
A
Then when it develops really powerful AI, it prevents other nations from doing the same. This is already beginning with export controls on China. And that kind of embeds its lead. And then it uses that AI to develop powerful new technologies and it's in control of those technologies. It uses AI to kind of automate cognitive labor throughout the US and maybe worldwide. And countries that don't use its AI systems will be really hard hit economically. And so we're kind of massively centralizing power in the U.S. and if the U.S. is able to maintain exclusive control over, over smarter than human AI, then it seems pretty plausible to me, very likely that the US would be able to rise to a strong majority, more than 90% of world GDP. And there's a few different dynamics that are driving that. First is that labor, currently human labor receives about half of world gdp. Just half of GDP is paid out on wages. AI will ultimately, and robots will ultimately be better than humans at kind of all economic tasks. And so if the US controls all of the AI companies that are replacing human labor, then that half of that kind of 50% of GDP which is currently going to human workers will ultimately be reallocated to paying to whoever controls and owns those AI systems, I.e. uS companies. There's a wrinkle there because some of that is physical labor. And US doesn't currently have a lead there. Physical robots. In fact, China is quite far ahead, but in terms of at least the cognitive aspects of our jobs. So we're talking significant fraction of GDP that would just now be reallocated to US companies that control AI. So that already gets them from 25% to above 50%. Then we've got this further dynamic, which is the dynamic of super exponential growth. So this relates to previous work I've done on how AI might affect the dynamics of economic growth, but kind of very potted summary is that it's often quoted that over the last 150 years, economic growth has been roughly exponential. And, and what that means is that if two countries are growing exponentially and one country starts off maybe twice as big as the other country, then at a later time still one country is twice as big as the other country. So let's say the US economy is 10 times as big as the UK economy. Then if they're both growing exponentially at the same pace, then 10 years later, again the US will still be 10 times as big as the UK. So that's exponential growth. That's what we've seen over the last 150 years. If you look back further in history, we see super exponential growth. That means that the growth rate itself gets faster over time. So an example would be that 100,000 years ago, the economy wasn't really growing at all. The extent it was growing, it was maybe doubling every 10,000 years or something in size, very extremely slow economic growth. Then going from about 10,000 years ago, it seems more like ballpark, there's a doubling of the economy every thousand years, still incredibly slow economic growth. You zoom back in and kind of 1400, you can begin to detect, okay, more like every 300 years or so, the economy is doubling. And then in recent times we've seen that the economy is doubling every 30 years. So essentially the growth rate is getting faster, the doubling times are getting shorter. That's super exponential growth. And there's various reasons, economic reasons, theoretical reasons, empirical reasons to think that AI and robotics, when it can replace humans entirely, will go back to that super exponential regime that has been at play throughout history. What that means is that growth is getting faster and faster over time. And the reason I'm saying all this, the reason this is irrelevant, is that go back to that example of the US and the UK. The US is currently 10 times bigger than the UK. If the US is on a super exponential growth trajectory, its growth is getting faster and faster over time. And that means that even if the UK is on that same super exponential growth trajectory, as they both go super exponentially, the US will pull further and further ahead of the UK because maybe the US is doubling in 10 years. Because it's already bigger, it's already further along the curve, whereas the UK is still doubling only every 20 years. And so that means that the US is now, rather than just 10 times bigger than the UK, the US is now going to be 20 times 30 times bigger in size in the UK. So if the US is able to, if there is super exponential growth and the US is able to kind of be bigger to begin with and therefore be further progressed along that super exponential growth trajectory, then that's another way that they could just continue to increase their size of the economic pie and ultimately come to completely dominate world gdp. So just to sum up everything I've said, today the US is 25% of world GDP. If it controls and develops AI that could easily boost it above 50%, I'd be very surprised if it didn't. And then from that point it's already bigger than the rest of the world combined. If it's able to then go on the superdex mental growth path, then it will go faster and faster over time and pull further and further ahead of the rest of the world. That may be able to grow super exponentially if they can also develop AI, but we'll still be falling further and further behind because of the nature of super exponential growth.
B
Yeah, this actually seems quite plausible to me and not very sci fi. The thing that seems quite sci fi is the notion that perhaps even one company could grow at such a speed that it would outgrow the rest of the world. How likely is that?
A
Yeah, great question. I think it's a lot harder. But it is surprisingly plausible. That first part of the argument I gave about how 50% of the world GDP is paid to human workers. If that went to AI, that'd be a big chunk. It is possible that one company could get a monopoly on kind of really advanced AI. So we already discussed some of the dynamics there, where again, the simplest one is just a combination of an intelligence explosion giving a company a big advantage, and then they're kind of buying up all the computer chips that the world is able to produce, outbidding everyone. If a company does that and already seems to be outbidding other companies on compute. Although Google also has a lot. If a company is able to do that, they could end up just one company in control of literally all of the world's cognitive labor, because human cognitive labor will at some point be kind of dwarfed by AI cognitive labor. So at that point that one company could be getting all of GDP, which is currently paid to kind of cognitive labor, which is a large part of the economy, as I said, maybe as high as 50%, but certainly as high as 30% of world GDP. If all that would then seemingly be going to this one company that controls the world's supply of cognitive labor. So though I think that would take time and obviously it's going to take a long time to automate all the different parts of the economy. There is just a basic dynamic by which one company can now be controlling double digit percentages of world gdp. And there's obviously questions, would a government allow that? Would they step in? And that's where we get into these dynamics of, well, this company has all these super intelligent AIs on its side. Maybe it's able to lobby, maybe it's able to do political capture to avoid the state setting in. Maybe it's able to be like, look, we're providing like economic abundance for everyone. If you step in, like, you know, that, that, that, that might not happen. You know, we're, we're underpinning your nation's, you know, economic and geopolitical strength. And if you try and you know, remove, you know, step in and nationalize, then, you know that's not going to happen. We're going to move to another country. So you can imagine maybe they convinced the head of state to kind of support them and there's some kind of alliance there. But it's not completely obvious that the company would be shut down. It would have certain types of serious bargaining power. And so if a company was able to maintain this position as kind of sole provider, cognitive labor, it would be able to get a, a significant fraction of world gdp. And then it's then possible that from there it could bootstrap. And this is where it gets a bit harder. But the tactic it would need to pursue is it already controls most of the cognitive labor, pretty much all of it. The thing it doesn't control is all the kind of physical machinery and all the raw materials that are also needed to create economic output. But it could pursue a tactic of kind of hoarding its cognitive labor so that no one else can ever have access to that and then kind of selling it at kind of really kind of monopolistic rents to the rest of the world because there's no one that can match it. It's offering everyone by far the best deal they can get, but just skimming off 90% of the value add from companies using its AI systems. So it's able to do that. Then it can kind of, it can kind of reap by far the majority of the benefits of trade, and then maybe can kind of increasingly buy up physical machinery and raw materials from the rest of the world, design its own robots, buy its own land. Imagine a kind of big special economic zone in Texas or something where this company is kind of unconstrained by kind of bureaucracy. And then it's also now got a big arm somewhere in Siberia and in Canada. It's kind of creating these big special economic zones by doing deals with specific governments. And I do think it's a bit of a stretch that this all goes ahead without various other powerful political actors pushing back. But the kind of basic economic growth dynamics are surprisingly compatible with a company ultimately coming to control most of the cognitive labor and most of the kind of physical infrastructure that its AIs have designed using all the parts that's bought from the rest of the economy.
B
Yeah. And do you think this is a risk factor for AI enabled coups then, just because you're concentrating all of the power and all of the resources into either perhaps one country or one company even?
A
Yes, I definitely do. The more realistic path is that a company kind of starts down this path of outgoing. The world gets kind of huge economic power, increasingly controls the country's industrial base, its kind of physical infrastructure, manufacturing capabilities, and then from there it's in a much stronger position to seize political control because it's got massive economic leverage. And then it can also increasingly gain military leverage because as it increasingly controls the country's broad industry and manufacturing, that will feed in to military power. So some of the possibilities I discussed earlier, where you could potentially have your AIs be secretly loyal, they're ultimately designed the military systems, or you could just instruct your AI systems to start making a military that is not legally sanctioned. But because the government doesn't have much to threaten you with, you get away with it. It gets a little bit tough. You probably need to do that in secret, otherwise the existing military could prevent it. But yes, I do think that being very rich helps with lobbying. It helps with all kinds of ways of seeking power. And then controlling a lot of industry can potentially give you military power.
B
You mentioned these special economic zones. That's one way in which companies could kind of bargain with states in order to have favorable regulation and to be able to carry out their projects without intervention. Basically another way for them would be to collaborate with non democracies that are perhaps controlled by a single, a small group or perhaps even a single person. And in that way it seems like perhaps it's easier to get Something done in a non democracy. And that is a way to grow fast. And so perhaps there are incentives for companies to place more resources in non democracies. What do you think about the prospect of non democracies out competing democracies when it comes to AI?
A
I think it's a really great question and it's tricky because I think I agree democracies have lots of checks and balances, they have a lot of bureaucracy, a lot of red tape and that will disincentivize AI companies for investing. And then additionally, if there are people really trying to seek a legitimate power will be easier to do in non democracies because they're less politically robust. So there are these various forces pushing towards this new supercharged economic technology being disproportionately deployed in non democracies. And I think that is scary. My own view is that probably we should democracy should, should kind of do everything they can to avoid that situation, make it much easier for AI and robotics companies to set up shop in democracies, remove the red tape, try and use export controls like are already happening to prevent technologies being deployed in non democratic countries. And that goes beyond China. There's obviously lots of countries that are not allied with China, but are also non democratic here. And the US is in a strong position because it does have this stranglehold on AI technology at the moment. So I do think it can be done. But yeah, in my view it will be really important to work very hard to find a kind of a non restrictive regulatory regime. And it will also be very important to really try and pursue innovative innovations within the democratic process itself, where democracy is great in many ways. It really distributes power and it has been very good at ensuring good outcomes for its citizens. But it's very slow and often kind of nonsensical because you have competing interests that are kind of stepping on each other's toes. And the resultant legislation is just a garbled mess. And so AI can potentially solve those problems. You can have AIs negotiating and thinking much more quickly on behalf of the kind of human stakeholders. You can have AIs nailing out agreements that aren't a garbled mess, but that really gave everyone what they truly wanted out of the legislation. And you can still do all of that really quickly so that you're not falling far behind the autocracies that have just got one person immediately saying what to do. And I think if we did that, democracies could outcompete autocracies. Because the big thing that often screws over autocracies is that one person is flawed, often makes big mistakes, people are afraid to stand up to them.
B
Yeah, that would be more of my assumption. I would assume here that perhaps democracies with market based economies have an advantage just because you can do kind of bottom up knowledge discovery. You can try different things out, you can see what works, you can have competition between companies and so on. And perhaps in non democracies, well, I mean you can have one person or a small group stake out a direction for what the country should do, but if that direction is wrong, it's probably difficult to change course.
A
Yes, I think you're completely right. I should have given more weight to that advantage of democracies in terms of the free market being, being in many ways much smarter. In terms of autocracies that are good at harnessing free market dynamics. My worry would be that AI helps them more than it helps democracies because AI will be able to kind of replace. Currently one person just can't think that hard, can't really figure out a good plan. But if that one powerful leader has access to loads of AI systems that can kind of think things through and investigate lots of different angles, then if they're following its advice, then they could get advice which lacks the flaws that today's systems had and they could potentially move much faster. But I think you're right, that kind of economic liberalism is still going to be important even after we get powerful AI systems and that could give democracy an advantage.
B
This is a bit of a tangent perhaps, but I'm thinking whether. So if you have a leader of a country that has a lot of power, perhaps complete power over that country, and that leader is equipped with AI advisors advising him and kind of laying out kind of the landscape of options for him to choose from, wouldn't his decision making still be in a sense bottlenecked by the fact that he's a human, by the fact that he has these flaws, that we all have the biases that we all have. So even with fantastic advice, I think it's quite plausible that he would still make the same mistakes that we see leaders make today.
A
I think that's true. I think it's also true in democracies, unfortunately, that if there's 10 negotiators and they each kind of still have biases and still refuse to listen to the wise advice they're getting from their AIs, that could still gum up the system. And yeah, it does depend on how much humans come to trust and defer to their AI advisors. There's a possible future where the AIs are just always nailing it. They're always explaining their reasoning really clearly and we are just increasingly convinced and happy to trust their judgment. If AI is aligned, I think that would be a great future because I do think humans have all these very big limitations and biases which if we can solve the alignment problem, AIs don't need to have. But there's also another future where humans just want to be the ones making the decisions, have these kind of pathetic motivations that they're still kind of influencing their decisions and that continues to limit the quality of decision making.
B
Seeing things from above, right from kind of like 10,000ft, how should we think about mitigating the risk of coups here? Is it, is it about removing people that would use AI to commit coups? Is it about kind of finding those people in the militaries, in the governments, in the companies, perhaps? Or do we have ways to reduce the returns to seizing power?
A
Yeah, I mean, from real 10,000ft up, the way I would characterize it is create a common understanding of the risks, build coalitions around preventing them, and then the existing balance of power can self propagate forward. It's in everyone's interest to prevent a coup. Currently no one small group has complete control or close to it. And so if everyone can be aware of these risks and aware of the steps towards them and kind of collectively ensuring that no one is going in that direction, then we can all kind of keep each other in check. So I do think in principle the problem is solvable and it doesn't require solving the risk of misalignment. Does require solving some tough technical problems. This doesn't. In the same way, yeah, you have.
B
A bunch of recommendations for mitigating the risks both for AI developers and governments. And perhaps we don't have to run through all of them, but you can talk about the most important ones for AI developers.
A
I might characterize this, I might kind of talk about it by going back to those three threat models we discussed earlier. So the first one was singular loyalties or overtly loyal AI systems, where again the main risk there is AI deployed by the head of state and the military and the government that's loyal to the head of state. And so the main countermeasure that currently appeals to me is for us to figure out rules of the road for these deployments. Obvious things like AI should follow the law. AI's deployed by the government shouldn't advance particular people's partisan Interests, but should only do official state functions. AIs in the military shouldn't be loyal to one person. Different groups of robots should be controlled by different people. And head of the chain of command can still be head of the chain of command via instructing other people that instruct those robots, but they shouldn't all go directly to head of chain of command because that centralizes military power too much. So fleshing out basic rules of the road of that kind and then building consensus around them. Because companies might want to say to governments, yeah, we don't want you to deploy our systems if they're willing to break the law, but if the government will have a lot of bargaining power, the executive in the United States can know it's hard for companies to stand up to them. So what we want to do is, you know, establish these rules of the road and then get broad buy in from Congress, from the judiciary, from, you know, other branches of the military, from many parts of the executive. So then it's very then hard for say, the President to say, yes, let's like make this robot army loyal to me. And everyone's like, obviously not. We've all agreed that makes no sense. You know, and then the President doesn't even bother trying because it's just clear that it would be a no go that, you know, their mind doesn't even go there.
B
In some sense. This is about kind of implementing the procedures and the transparency rules that we know from democracies today into how we use AI, both in governments and in companies?
A
I think exactly, yeah.
B
Do you worry here that when. So the government is looking at these companies from the outside and they don't have full insight into what's going on. There are kind of protections for private companies that mean that they can, they can do things in secret without the government knowing, at least as things stand. Now, is that something that would evade these mitigations you're. You're thinking of?
A
So I mean this, for this first bucket, the singularities bucket, it's mostly the, the kind of heads of state that I would be worried about. So it actually is probably good for the government, or at least for the head of state themselves, not to have full insight into literally everything the company is doing, because that would give them too much power. But actually having different parts of the government having insight into what the lab's doing, I think is very good. I'm a big fan of transparency. And we do have a good set of government checks and balances from different government bodies that we can deploy to kind of keep the lab in check using these other bodies, but also not allow the executive branch and the President to get excessively powerful. So that's the mitigations, the kind of singular vert loyalties. In terms of secret loyalties, the key mitigation is what I'm increasingly calling system integrity. That is using established cybersecurity practices and machine learning security practices, that is kind of preventing sleeper agents and backdoors in machine learning models, using all of that to ensure that your development process for AIs is secure and robust and that no malicious actor, be they an employee in the post training team at a lab, or be they the CEO of the lab that is either malicious or is being threatened by the Chinese government to kind of tamper with model development, that no person or no small group is able to significantly tamper with the behavior of AI models. And no group is able to get illegitimate access to AIs that would help them seize power. So that's this idea of system integrity, which is essentially a technical project which does just draw on existing practices but is not yet implemented in any of the top labs. I will quickly shout out for non lab, for people listening that aren't working at labs. I think there's a lot of really good technical research that could be done.
B
On.
A
Investigating the conditions under which you can insert a steeper agent without a defense team knowing. And there's just loads of research that can be done in terms of the different settings there for attackers and defenders, which could then inform what parameters we need to be in place to achieve system integrity. If it turns out that it's very hard to make a sleeper agent except in the final stage of training, that's really useful to know because then we can focus our efforts within labs at that final stage, just as a hypothetical example. So that's the kind of key mitigation in my mind for the secret loyalties. And then I'll quickly do for exclusive access.
B
That one seems more difficult. I don't know just from me reading and preparing for this interview, that one seems like a difficult one to handle. Where this is in some sense a deep trend in history and in, in the kind of history of modern economics that you do see faster growth rates and you do see concentration into bigger and bigger economies, both in countries and in companies. So are you in some sense pushing against underlying trends if you're trying to mitigate exclusive access to advanced AI from one actor?
A
I think you can, you can do this in other ways. So you can have the law require that AI share that AI labs share their powerful capabilities with kind of other organizations to act as a check and balance. So labs should share their R and D capability, AI R&D capabilities with evals organizations.
B
Here you're thinking about giving insight into what they're capable of, not not actually sharing those capabilities. That would be too big of an ask. I think.
A
I do mean API access. So if a lot of the work in developing and evaluating systems is now done by AIs, then we want an evaluation organization like Apollo or Meter to also be uplifted. And so we want them to have access to really powerful AI that can similarly stress test how dangerous the frontier systems are. If they're only using human workers, then that's going to be a big disadvantage. So no, I do want API access to powerful capabilities for other actors. For example, cybersecurity teams in the government and in the military should have access to the lab's best cyber capabilities. And again, that should be a requirement by law. So generally, even if there's a natural tendency towards centralization of power in one organization, you can still require that that organization share its systems with the checks and balances. That's one thing. And the other thing is kind of preventing anyone at this organization from misusing the powerful AI systems. So the, the biggest thing on my mind here is that today we still have helpful only AI systems where you can kind of get access to the system and then it will just do whatever you want, no holds barred. I don't think there should be any AI systems like that. I think you should always have at least a classifier on top of the system which is looking for harmful activities and then kind of shutting down the interaction if something harmful is detected. And if you have a special reason to use cyber offense for your job, or you have a special reason to do potentially dangerous biology research, you can have that classifier allow certain types of activity, but you should never have anyone accessing a system where anything is allowed. No one has legitimate reason to access an AI that will literally do anything. So what I want to aim for is a world where, yes, if there's a specific reason why you need to use a dangerous capability, absolutely you can use that system. But that system will just do that one dangerous domain. It won't kind of do anything you wanted because that, you know, that's a very scary situation where there's, you know, there's a hundred reasons why the CEO could ask for access to a helpful only system. You know, maybe the guardrails are annoying, maybe he wants to kind of, you know, do something which the model is reluctant to do. But today when you asked to remove some guardrails, you're removing all of the guardrails and now there's no holds barred. So instead we should be flexibly adjusting what guardrails are there by the use case and just never have a situation where there's no guardrails. I think that could go a long way towards helping if that was robustly implemented.
B
With all of these mitigations for both secret loyalties and exclusive access and singular loyalties, you would worry that they would be disabled by the group planning a coup, right? Say that, for example, you are the CEO of an AI company and you're giving API access to evaluations, organizations, testing your model, trying to see what they're capable of. Maybe you just cut off access before you get to the really powerful model that could actually be the model that helps you conduct a couple of do we have ways of making sure these mitigations are entrenched in such a way that they can't be removed by the group planning a coup?
A
This is a great question. It is pretty tricky. CEOs by default have a lot of control over their organizations and similarly heads of state, including US President has a lot of control over the military and over the government. So yes, there's a risk that one of these powerful individuals realizes that maybe they want more influence by gaining control over AI and notices that there's these kind of pesky little processes that prevent that. And it's like, okay, let's remove them. I can give easy, say, productivity reasons to prevent them, red tape reasons, and if they can make a plausible argument, then it could be hard to oppose them. So I do think it's a big issue. But I'd say a few things. Firstly, something I mentioned earlier, I don't think that anyone is today planning to do an AI enabled coup. The way I think this works is that people are faced with their kind of immediate local situation, something they want to do over the next month, and the blockers that they're facing to doing that specific thing. And what tends to happen is people tend to want more influence that helps them get stuff done. And so people will kind of bit by bit kind of move in the direction of getting more control over AI, but they won't be kind of thinking, yes, I need to make sure that I remove this whole process because that will allow me to do an AI enable queue. That's kind of unrealistically galaxy brained. And so what we could do is we just set up a very efficiently implemented and very reasonable set of mitigations that doesn't really prevent CEOs from doing what they're trying to do. And so the CEO doesn't find in their day to day that they're wanting to kind of remove these things that are holding them back. But because these mitigations are here, the CEO never gets to a place where they're anywhere close to being able to do a couple of, or where there's any kind of pathway in their mind to be able to doing a coup, because they're constantly prevented from getting access to kind of really powerful AI advice that might point out ways in which they could do this because they're surrounded by colleagues that strongly believe that these mitigations are sensible and reasonable and in fact they are well implemented and there aren't many downsides. Maybe an environment where they kind of get kudos for the fact that they've said, yep, obviously I'm not going to get access to helpful only systems. That's crazy. And then that's kind of like something, something that, that, that makes them seem good. So that, that, that's one thing to say. Another thing is again, going back to this point, that there are currently checks and balances and there is not currently a situation where one person has power. If the entire board of a company and other senior engineers recognize the importance of the mitigations, know about this threat model, then they will notice if the CEO is moving that direction. And similarly within the government, there are checks and balances and they could be activated if people are looking out for it.
B
Do you think these traditional oversight mechanisms, like a board being in control of the CEO, being able to fire the CEO, or the possibility of Congress or the Supreme Court kind of overruling or constraining the U.S. president, do you think those will persist in environments where AI is moving very fast and AI capabilities are growing at a rapid pace?
A
It's a great question. Here's one story for optimism. Today things are moving fairly fast, but those checks and balances are somewhat adequate, at least to preventing really egregious situations. By the time the AI is moving really quickly, we'll have handed off a lot of the implementation of government, the implementation of things in the AI companies, the research process will have handed it off to AI systems. And when we do that handoff, we could program those AIs to maintain a balance of power. So rather than handing off to AIs that just follow the CEOs commands or AIs that follow presence commands we can hand off to AIs that follow the law, follow the company rules, report any suspicious activity to various powerful human stakeholders. And then by the time things are going really fast, we've already kind of got this whole layer of AI that is maintaining balance of power. The whole AI government bureaucracy, the whole AI kind of company workforce, they are better than humans today at standing up to misuse. Potentially they are less easily cowed and intimidated and they could actually make it harder for someone in a position of formal power to kind of gain excessive influence. So this is like the flip side of the singular loyalties where you potentially deploy these AIs that are explicitly loyal. You can actually kind of instead get kind of singular law following and balance of power, maintaining AIs that you deploy. And so the hope is that by the time we really things are beginning to go kind of crazy and we're really seeing speed ups from AI, we've already kind of set ourselves up in an amazing way to maintain balance of power. And there's this critical juncture where we are handing off to AIs and it's just what are those AIs, what are their loyalties? What are their goals? And I think we can gain a lot by making sure that those AI systems are maintaining balance of power, reporting illegitimate suspicious activities, and are not overly lodged to any one person.
B
How do you think the risk of AI enabled coups interface with kind of more traditional notions of AI takeover? So just a misaligned highly capable or advanced AI system taken over, kind of contrary to the wishes of the developers or the governments?
A
Yeah, I mean there's some close analogies. Perhaps the most analogous case is the case of secret loyalties where you've got these AIs that have been told by the CEO to have the secret goal of seizing control and then handing control to the CEO. That's just very similar to AIs that wanted to seize power for themselves secretly. And all the same stories could apply where the aisle of make military systems and then they control the military systems and the robot army and then they seize power. The only difference is were they seeking power because it just kind of accidentally emerged from the training process, which is the misalignment worry, or were they seeking power because the CEO programmed them in that way? But that's the kind of seed of the power seeking. But then with the secret loyalty threat model, the rest of the story is pretty similar. There's still differences. In the secret loyalties case, the CEO might be doing more to help the AI'S along with their plan. Maybe even in the misalignment case, the AIs have managed to kind of manipulate the CEO into doing similar things. So that's the case where it's most analogous. I think another difference that's salient to me is that if there are lots of different AI projects, then an AI enabled coup seems a lot harder because you need lots of different humans to kind of coordinate, to kind of seize power together. Which seems, while I can totally believe that one person might try and seize power, does seem less likely to me that there'd be loads and loads of humans that would want to do that from lots of different labs. Whereas from the misalignment story, it is more likely the case that if one of these labs has misaligned AI, then maybe lots of them have misaligned AI. And so then it's more likely that you would have maybe 10 different AIs colluding and then seizing power and taking over. And so that kind of collusion between multiple different AIs is more likely in the case of misalignment than in the case of.
B
And I enable Q just because if there's one misaligned AI, then there's something about the training process for AI systems that are causing misalignment, and then it will be a common feature among many companies.
A
Exactly. Whereas just the fact that one CEO instructed a secret loyalty would not to the same extent make you expect that other CEOs have done the same.
B
So you mentioned this possibility, but do you. You think it's. Yeah. What do you think of the prospect of a president or a CEO of a company being duped by misaligned AI into conducting a coup on its behalf. So you can imagine a president or a CEO kind of thinking that he's conducting a coup to remain in control, but he's actually acting on behalf of a misaligned AI.
A
Yeah, I think it's an interesting threat model and some people who think about AI takeover threat models take it pretty seriously. And it's just a case where we're just completely mixing these two threat models together. People who are worried about AI takeover for this reason should be very supportive of the kind of anti coup mitigations I'm suggesting, because if we implement checks and balances that prevent any one person from getting loads of power, then that AI will not be able to convince them to try because they just won't be able to succeed. So I see this as an additional reason to worry about AI enabled human coups and to try and prevent them is that yes, even if no human wants to do this, normally misaligned AI might make them try. In terms of how plausible I find the threat model, honestly, I think that if a human tries to seize power, the main reason is that human wanted power. This is just something we know about people, we know it about heads of state today. It's very clear that many heads of state in the most powerful countries in the world are very power seeking. We know it about CEOs of big tech companies. We know about some of those leading AI companies that we do know that they're very power seeking, those CEOs. And so I don't think we need to theorize that they were massively manipulated by the AI and convinced to become power seeking. I think it's more likely that if they seek power, they just did it for the normal human reason. I do think AI will ultimately get good at persuasion. I don't particularly expect it to be hypnotic level persuasion though. Obviously there's massive uncertainty here. And yeah, I do think that a very smart AI, where there's a human that's already kind of interested in seizing power and it already kind of makes sense for them to maybe do it and a misunderstand AI could totally nudge them in that direction and then could implement that in a way that actually allows the AI to seize power later. I think that is very plausible.
B
When we're thinking about distributing power and kind of having this balance of power, we can imagine the models being set up via post training, via the model spec, via various mechanisms to obey the user, unless what the user instructed to do is in conflict with what the company is interested in, and perhaps obey the company, unless what the company is using the model for is contrary to what the government kind of permits. But when we set it up in those levels, you ultimately end up with the government in control in some sense. And I guess that exposes you to risk of a government coup then if you have at the ultimate top layer of the stack, here's what the models can and cannot do according to the government.
A
Well, I'd say a couple of things. First is that the government isn't a monolithic entity. And so that government decision of all the balance changes be could be informed by multiple different stakeholder groups. And then ideally it's ultimately democratically accountable. I do think that democratic accountability becomes more complicated in a world where there's massive change in a four year period.
B
Just for the simple reason that there's no election during a period where a massive change is happening. So the feedback loop is too slow.
A
Exactly. I think the risks of AI enabled coups will probably emerge and then be decided within a four year period. As in like it will be resolved whether or not it happens or doesn't, all without any intermediate election feedback. That doesn't mean that democracy can't have an effect because politicians anticipate what future elections will find and want to maintain favor throughout their terms. But it does pose a challenge. But sorry, I was kind of saying, even absent that there's many different stakeholders in the government and so it would have to be a large group of government employees that were kind of trying to do a coup and then they would have kind of. The companies would know that they were saying these odd restrictions on the kind of behavior and so the companies would know and they have leverage and power and then it could go public. So I don't think it would be that easy for the government to coup.
B
Perhaps there's a difference also between allowing the government to set restrictions on what the models can do and then allowing the government some kind of access to commanding the future AI systems in certain directions. So it's kind of setting limits versus steering the systems.
A
Yeah, exactly. I mean the distinction I was going to highlight was between specifically making AI systems loyal to for example, the head of state. And they're setting very broad limits where there's just like you can pretty much do whatever you want except for these obviously bad things where that second option doesn't really enable anyone to do a coup, it just enables everyone to do whatever they want. And then you've kind of blocked out all of the kind of coup enabling possibilities through those limits as long as you haven't made those systems loyal to a small group. So given that there's this obvious option to just put in these limits that block coups but don't enable coups. And given that there's a wide range of stakeholders that could potentially feed into what the AI's kind of limitations and instructions are. I think it's very, very feasible to get to a world where there's robustly not centralization of power. There's obviously a big uncertainty over whether we will actually get our act together and get those limits put in place in the right way.
B
When do you think the threat of AI enabled coups materialize? Is it at some specific point in AI capabilities or does it simply scale with the systems getting more advanced? When do you think the threat is at its peak?
A
It's a good question for the threat models that I've primarily focused on, they require pretty intense capabilities. So that, for example, the secret loyalties threat model more or less requires AI to do the majority of AI research. So we're talking about fully replacing the world's smartest people in a very wide range of research tasks and coding. That's pretty intense. And then a lot of the threat models that I focus on route through military automation, that is AI and robots that can kind of match human boots on the ground. And that's pretty advanced. Again, that said, I think you can probably do it with less advanced capabilities than that. So drones today are already pretty good, already providing, making a big difference in some military situations. So it's not out of the question that more limited forms of AI and robot military technology could be enough to facilitate a coup. It's a bit harder because if they're limited, then there's a question of why the existing military doesn't just kind of seize back control after a bit of time. And so probably that scenario also has to involve things like maybe the current president supporting the coup and therefore pressuring the military not to intervene or some other source of legitimacy for the coup beyond the kind of the AI controlled drones. And then there's also kind of more kind of typical types of backsliding like has already been happening in the US that I think could be exacerbated through AI enabled surveillance and AI kind of increasing state capacity in other ways. And again, that backsliding doesn't require super powerful AI. You could probably do a lot of monitoring, a lot of kind of content moderation on the Internet, a lot of surveillance. With today's systems, it doesn't get you all the way to one person having complete control where they can just quash any resistance with a robot army and replace everyone in their job with an AI and so no one has any leverage. So I think to get to that real intense, this is the most intense form of concentration of power via AI that requires really powerful AI, but to just kind of significantly exacerbate existing trends in political backsliding and to make it easier to do a military coup. I think more limited systems would suffice.
B
Yeah, we discussed earlier the possibility of one country or one company outgrowing the rest of the world and kind of concentrating power into those entities. Now you mentioned one person. Do you think that's actually a plausible scenario in which you have say, one CEO of one company being the person in control of the world via a concentration of power and then a coup?
A
100%? Yeah. I mean the story I told earlier about secret loyalties, meaning that now we backdoored wide range of military systems, meaning that you can seize power. That's one route. And then again there's this other, the company masses amounts of economic power by kind of having a monopoly on AI cognitive labor and then leveraging that to get more economic power, more political influence. Yeah, I do think it's possible. Again there's this big shift once AI can fully replace humans, where today one person can never have absolute power. They have to rely on on others to implement their will.
B
And this is what makes currently existing dictatorships unstable, where there's always a threat of kind of internal revolt or outside factors threatening the dictatorship. But this could potentially change.
A
Yeah, there's always a threat of revolt and then to guard against that threat, the dictator needs to share their power to some extent has to compromise. But yeah, you could get it all concentrated in one person. With sufficiently powerful AI, do you think.
B
We move through a period of increased threat of AI enabled coups and then reach some kind of stable state? Or do you imagine that there's a constant kind of risk of AI enabled coups in the future?
A
I think we move through it, yeah. It's this point about once we have deployed AI across the whole economy, the government, the military, and if those AIs are maintaining balance of power, then we could fully eliminate the risk of AI enabled coup. It'd just be as if our whole population was just so committed to democracy, would never seek power, never help anyone else who wanted to undermine any democratic institution. We already have strong norms favoring democracy, but they're far from perfect and they have been eroded over recent decades. But you could just get rock solid norms. They're programmed in, they cannot be removed except by the will of the people. I mean, there is a bit of a question because you still want to give the human population the ability to change the AI's behavior and its rules. So the human population could always choose to move to an autocracy. So I suppose I shouldn't say that we could fully eliminate the risk because we will always have that democracy. There's always this point that democracy could vote to stop being a democracy. But I do think we could get to a point where it absolutely cannot happen without most people wanting it to happen.
B
So you would get to a point in which future AI enhanced societies you could say are more stable than current democracies and they're less at risk of coups than or democratic backsliding than current.
A
Democracies Much, much more. Yeah, you could get Much more robustness there. I mean, there's this constant dynamic in today's societies where people care about democracy, but they also care about a host of other things, their own achievements, various other ideological commitments. And so depending on how dynamics play out, depending on how technology evolves and what people's incentives are, sometimes people push against democracy. That's what the Republican Party's been doing. In some ways that's what Democratic Party has done as it's increasingly put pretty ideological people in powerful institutions. So with AI, you get much more control over those dynamics because you can just make it much more the case that democracy is not being compromised.
B
Are there any ways for us to. Are there any kind of risk factors we can look at if we are interested in predicting coups? Do you think there's something we can measure or something we can track to see whether we are at risk of an AI enabled coup?
A
It's a great question. I don't think I have an amazing answer. But some things that come to mind, the capabilities gap between top AI labs and then the gap again with open source, the degree to which AI companies are sharing their capabilities with the public, and if not with the public, then with multiple other trusted institutions. You know, like sharing their strategy capabilities with kind of US political parties and parts of government. The extent of economic concentration, you know, how much, what are the revenues and net worth of particular companies, particularly AI companies. Another one, what is the extent of government automation and military automation by AI systems and when that automation is happening, how robust are the guardrails against breaking the law and guardrails against other forms of illegitimate power seeking? How much transparency does the public or the judiciary or the Congress have into how dangerous AI capabilities are being used by AI companies and by the executive branch. So, you know, take the example of military R and D capabilities. That is really smart AIs that can design super powerful weapons. It's scary if companies can just use those military R and D capabilities without anyone knowing. It's also scary if a small group of people from the executive branch can use those capabilities without anyone else knowing how they're using them, because they could be designing powerful weapons and making them loyal to a small group. So transparency into these high stakes capabilities and how they're being used by a broad group, it doesn't have to be public, probably shouldn't be public, but we have checks and balances already. So another kind of question is, as these high stakes use cases start occurring or they become possible, do we know that there's transparency requirements in place as we increasingly see AI companies contracting with Palantir and other military contractors, we can kind of begin to see that they're making increasingly powerful weapons. Is there a process of oversight? Do we know that if someone was trying to make AI military systems the law to them that it would be spotted? That's another indicator. You can look at all the standard democratic resilience indicators that the social scientists have come up with. There's various things about free and fair elections, about civil society, about freedom of press that have been getting worse recently in the U.S. but there's various indicators here. You can look at the degree of government censorship over freedom of speech or what's on the Internet, and the degree of surveillance that the government's doing.
B
If you take all of these things into account, how do you think about the risk of an AI enabled coup in the next 30 years?
A
Say next 30 years? I think it's high. I think the risk is high. I would guess it's 10% or something. And that to be clear, if it was just existing political trends ignoring AI, I would be maybe a few percentage, maybe on 2% or something. There's definitely a risk of that. And I'm thinking about the US here. A big part of my current worries are not about the indicators, but it's about my expectation that AI capabilities will keep increasing quickly and even more quickly. And then the kind of absolute lack of interest in regulating AI companies right now in the US and the difficulty that we will have of constraining the executive under the current situation where the President is using sophisticated legal strategies to increase their own power and is succeeding on many fronts. The US is not doing a great job at constraining the executive. So companies are unconstrained, the executive is poorly constrained. Those are the key threat actors here. So with fast AI capabilities progress, plus that lack of constraint, lack of transparency, the default, is that a lot of those indicators, I said get worse and none of the indicators get better, like transparency. And so that makes me think this is, this is very plausible.
B
Yeah, I mentioned 30 years, but what about five years?
A
Five years? That's tough, isn't it? It's really tough. I mean, yeah, I think there's a risk. I wouldn't think there was a risk if it wasn't for the AI research causing an intelligence explosion angle. But AIs are a lot better at coding and cognitive kind of research related tasks than they are at, for example, controlling robots and stuff. And so even if the threat model ultimately comes through robots or comes through crazy levels of Persuasion. It's just, you really can't rule out a scenario where AI research is automated in three years time. Then in four years time we've got super intelligent AI controlled by a few, few people. Maybe it's got secret loyalties, maybe it's being deployed in the government and being overtly loyal to the President. And then a year later it's backsliding or it's political capture or it's robot soldiers.
B
Yeah. How do you think about the badness of the outcomes here? How much does the badness depend on the ideologies of the people who are, who are conducting the coup? Or what should we look out for? Because I mean, I guess we can rank coups by badness, which is not an exercise I think we should actually attempt, but we can kind of talk about the factors involved, about what would be the worst kind of coup and what would be a slightly better kind of coup, slightly less bad kind of coup.
A
Yeah. So we could, you know, let's imagine it's one person that sees power. Actually that's the first distinction to draw. If there's a group, then even 10 people is better than one person.
B
And why is that?
A
Yeah, so 10 people, you get a diversity of perspectives, so more kind of moral views represented and there's more kind of room for compromise between those perspectives. There's more room for kind of reasonable positions to win out as there's kind of some, some deliberation as, as actions are decided upon. There's slightly less intense selection for psychopaths than if it was just one person. So yeah, if it's just one person that's bad, that's particularly bad. 10 people still very bad, you know, 100 people still pretty bad. But you know, it's that there's big differences there, big differences. If we're now just thinking about one person or the average person in the group, then we could think about how competent they are and then we could say something about how virtuous their motivations are. Well, I do think competency is important. I think it's probably underrated in most political discussions how important it is to just be really competent. Thinking about something like responding to Covid or thinking about something like trying to de escalate a conflict, Russia, Ukraine or trying to de escalate Israel conflict. Actually just being very competent and very good at getting things done is important. And as we mentioned, if you're just willing to rely on AIs and you align those AIs in the right way, anyone could be really competent. But that's not guaranteed. People may really want to cling to their current views without changing their mind. Let's take the example of Donald Trump. If a really smart AI system told him, look, tariffs are definitely bad for the US Economy, they're definitely bad and won't give you what you want, would he change his mind? I. I would guess no. So lots of smart people have already been saying that. And him and his supporters. I don't actually know the economic details here, but my understanding is that most people think that they're pretty bad. And it'll still be the case that Trump will be able to find people telling him that what he thinks is good, and he'll be able to program his AIs to keep telling him that if he wants to. So there's no guarantee that they. That he will become super competent or that whoever sees power comes super competent.
B
So there's this kind of like, there's a form of loyalty that actually undermines competence just because you're loyal to such an extent that you're not providing feedback that's useful because negative feedback feels bad to receive. And so there's that kind of loyalty. I mean, maybe this is a bit contrived, but do you think there's a sense in which, in the singular loyalty scenarios, the AIs could be so loyal that they are. They're kind of undermining the competence of the person that they're singularly loyal to?
A
Yeah, it's a really great question. I haven't thought about this, but, yeah, in a way, the most extreme version of singular loyalties will just agree with whatever the most recent thing that the dictator has said. It's a version of sycophancy which we already see without questioning. And we'll do that even when it's not in that person's interests, because that's the kind of type of loyalty that's demanded where there's a more kind of sophisticated type of loyalty where you're still completely loyal, but you are also willing to challenge them when you think it's in their best interests. So that's a really nice distinction. And yeah, I suppose one way of thinking about competence is thinking about what kinds of loyalties the dictator would demand from their AI systems. Another way of thinking about it is how much they would listen to the AI advisor. Even if the AI has the kind of sophisticated type of loyalty and it's trying to tell the dictator what to do, the dictator could just ignore them. And you see that again, AIs are fairly sycophantic they will also challenge you sometimes and then it's up to you whether you listen. So that's all the confidence bucket, which I think is really important. And I do think there are differences between potential coup instigators on that front, which could be significant. Where yeah, I guess my expectation would be that kind of lab CEO coups would, would be more competent than heads of state. But Even within lab CEOs there are some that are more dogmatic than others. And I think that dogma would get in the way of competence. That's competence. And the other thing I mentioned was kind of broadly what are your goals, what are your values? Or more character. And here one thing I think is really important is being open minded.
B
Being.
A
Willing to bring in lots of different diverse perspectives into the discussion and empower them to really represent themselves and grow and flourish. I think a very bad thing would be a particular person becomes dictator, they implement their vision for society. End of much better would be empower all the different kind of ideologies and ideas to kind of become the best versions of themselves and then we can kind of collectively grow and improve our understanding of how to run society. So sometimes people focus when they're thinking about values on like okay, are you this type of utilitarian or oh no, I hope you're not a deontologist or it can get very kind of specific and finger pointing. You know, my view is more that, you know, we don't really know what the right answer is. And the most important thing is is is being pluralistic and you know, letting a thousand flowers bloom.
B
So we, we discussed the possibility of getting to stable state in which we've avoided an AI enabled coup and now we have say we have aligned superintelligence kind of that where the risk of coup is, is very low. Do you think this is something that happens for one country and then that one country is in control of the world to such an extent that this is not a process that other countries are undergoing. To be more concrete here, for example, if the US goes through a risk of AI enabled coups but manage to kind of stay to remain a stable democracy, is it the case that Russia or China will go through a similar period of risk of coups?
A
It's a great question and it will depend on the US's posture towards the rest of the world geopolitically. And it will also depend on whether the US has gained a huge military and economic advantage like outgrowing the world or just developing powerful military technology as we were discussing previously. But you can imagine one scenario where the US isn't that much more powerful than the rest of the world yet, and isn't that inclined to intervene, which has been kind of the recent trend. And then China developed some really powerful AI a few years later and Xi Jinping uses it to cement his control over China. So then you now have one kind of AI enabled dictatorship that is extremely robust. And then you have the kind of US which has avoided that risk. And now they're kind of, maybe they're competing against each other and kind of the Cold War iii. I'm trying to kind of. Cold War ii, sorry, trying to outgrow the world. Or maybe they're striking deals because they recognize it's not good to compete and China just kind of indefinitely remains a dictatorship and that's just a permanent loss for the world. But you could also imagine a different scenario where the US is very far ahead and maybe, you know, it just wants to really secure its position geopolitically. And so it, you know, it instigates enabled coups in other nations where it's really putting kind of US representatives up on top of those, those nations that could be through secret loyalties. It could sell systems to, you know, sell, sell AI systems, let's say to, to India that are secretly loyal to, to U.S. interests. Or it could give some particular politicians in India access, exclusive access to superintelligent AI to help them gain power. So you could apply those same threat models we've discussed, but with the kind of US pulling the strings. Or you could have the US just kind of taking control of other nations in more traditional ways, just military conquest and kind of really leaning heavily on kind of extracting economic value out of other countries as they outgrow the world. So yeah, kind of wide range of options here really.
B
Yeah. As a final topic here, perhaps we can talk about what listeners can do if they want to help try to prevent AI enabled coups and specifically where to position themselves. Should they be in AI companies? Should they be in governments? Should they be in perhaps eval organizations? Where is the position of most leverage?
A
Great question. I think being a lab is a great place to be. I've talked about system integrity, kind of robustly ensuring that AIs don't have secret loyalties and behaviors intended. That's something that companies need to implement. So if you have interest or expertise in sleeper agents or backdoors to AI models or cybersecurity, then I think being part of a lab and helping them achieve system integrity is an amazing way to reduce this risk. Another thing you can do at labs, if you're interested in, if you're worried about the risk of heads of state deploying loyal AIs and seizing power, you can help labs develop terms of service where when they sell their AI systems to governments, they have certain mitigations against misuse. Maybe one way to frame this is, look, you're using really powerful AIs and we can't guarantee the safety of those AI systems unless we have some degree of monitoring to ensure that the AI systems aren't doing anything unintended. That monitoring could then be sufficient to allow for the prevention of coups. Because you'll be monitoring not only for kind of accidental misaligned AI behavior, but that will also thereby mean you're monitoring for a bad human actor giving them illegal instructions. So labs will be drawing up contracts with governments terms of service. They will be thinking about the guardrails, if any, that are placed on the systems that they sell to governments. But I think there's very careful work to be done thinking through, okay, how can we structure those guardrails? How can we explain them in a way which is very unarguable? And it doesn't seem like we're kind of trying to constrain the government. Private companies don't. It's not really legitimate for them to kind of constrain the government. But I do think there's an important thing to be done here in preventing air enabled coups. So kind of threading that needle. There's another thing you could do in a government, in a lab. But you could also do that kind of work for a think tank or for a research organization that's kind of interlinked with a government like Random, I think could potentially do some of this kind of work. Thinking about what should be in the terms of services between labs and governments. Let me think. I think another big thing is that for this system integrity, yes, we want labs to implement it, but we also want there to be some external organization that can certify that currently no external organization is working on this. Meter's not working on it, Apollo's not working on it. I don't think any evaluations organization is. But there will be at one point in time a very strong demand for guarantees that the systems that these companies have developed don't have super agents. Because if you're deploying a system in critical infrastructure, in important government roles in the military, you want to know that it does what it says on the tin. And so I think that a company can make a wise investment by developing that capability. The ability to understand how to insert secret loyalties, how to detect them. Being able to assess a company's process to see whether it's sufficient.
B
How close would that work be to kind of interpretability work? Isn't there overlap there? Kind of like looking into what is the system actually doing? Fine. Kind of reverse engineering some process to see how things work. Is detecting sleeper agent like a specific constraint problem?
A
Yeah, it's a great question. It's a specific type of problem. So you could use internals work like interpretability to detect steeper agents. That's not the only method you could use. You could use behavioral tests, you could look at the training data. I think a mixture of those is useful. It's a specific adversarial setup. So interpretability, we're just often interested in all kinds of kind of facts about the model's behaviour and its cognition. With secret loyalties, we're specifically concerned with one red team trying to produce a secret loyalty and then a blue team trying to detect it. And so it's adversarial in that the blue team has a tough task. The red team will be purposely trying to hide the sleeper agent. And so that will raise a whole host of new considerations and important things to dig into. So you can imagine being like allowing the red team full control over all aspects of the development process. Can they create a secret loyalty? Maybe they can. Maybe. It's pretty hard to test. Okay, but now you only allow them to intervene at one part of the development process, or you allow the blue team to look at all the data that they use for training, or you allow the blue team to look at all of the code that they wrote and you could get potentially quite a sophisticated understanding of the conditions under which the blue team win. Conditions under which the red team win. This research doesn't need to be done in a lab. It could be done by an external organization. I think it's big, big missing focus of today's technical work. And ultimately that would inform the assessments of the lab's attempts to do system integrity. So for technical researchers out there, I'd really highlight that possibility. Another kind of piece of work for the right person would be beginning to understand the kind of existing military thinking around autonomous systems. This is already obviously a live issue for militaries. They are increasingly deploying AI. It would be nice to marry up that existing expertise with these kind of risks about more powerful systems enabling coups and kind of get to a consensus within that military community of basic principles like law following, like distributed control over military systems, and figure out a kind of a military procurement process which is both practical but also robustly prevents this kind of stuff. So if there's anyone listening that has a way in, I think that's potentially pretty valuable. Although there's also risk of poisoning the well if it's done badly. So proceed with some care.
B
Yeah, perfect. Thanks for chatting with me, Tom. It's been great.
A
Yeah, real pleasure. Thanks so much, Gus.
Podcast: Future of Life Institute Podcast
Date: July 17, 2025
Host: Gus Docker (B)
Guest: Tom Davidson (A), Senior Research Fellow at Forethought
In this episode, Gus Docker interviews Tom Davidson about the potential for advanced AI systems to enable or facilitate the overthrow of governments—a phenomenon they term "AI-enabled coups." Rather than focusing on "rogue AI" scenarios, the discussion centers on how powerful actors (such as state leaders or corporate CEOs) might leverage AI to undermine democratic institutions, seize power, or entrench authoritarian control. Davidson explains various threat models, explores historical analogies, outlines concentrations of power risks, and imagines mitigation strategies to prevent such scenarios. The conversation provides both concrete hypothetical examples and high-level frameworks for thinking about future AI risks in politics, economics, and military affairs.
"What I've been focused on recently is not the kind of traditional idea that AIs themselves will rise up... but that a few very powerful individuals will use AI to seize a legitimate power for themselves." – Tom Davidson [01:29]
"There is a bit of a phase shift at the point in which AI can fully replace other humans in government and the military... a leader doesn't need to rely on anyone else." – Tom Davidson [09:26]
"Sleeper agent is the standard term used in the technical literature... the kind of secret loyalties is just what I call the scary situation where you now have a... Sleeper agent which is specifically loyal to one person trying to help them seize power." – Tom Davidson [17:20]
"If a lot of the work in developing and evaluating systems is now done by AIs, then we want an evaluation organization like Apollo or Meter to also be uplifted." – Tom Davidson [61:47]
"All these kind of standard tools where it's now a lot harder to point at one thing that's clearly egregious. But when you add up hundreds of little paper cuts for democracy that are systematically administered, you're seeing a real kind of loss of democratic control." – Tom Davidson [12:15]
"I do think that being very rich helps with lobbying. It helps with all kinds of ways of seeking power. And then controlling a lot of industry can potentially give you military power." – Tom Davidson [44:40]
"...probably we should—democracy—do everything they can to avoid that situation, make it much easier for AI and robotics companies to set up shop in democracies, remove the red tape..." – Tom Davidson [47:32]
"The key mitigation is what I'm increasingly calling system integrity. That is, using established cybersecurity practices and machine learning security practices... to ensure that your development process for AIs is secure and robust..." – Tom Davidson [57:39]
"We could program those AIs to maintain a balance of power. So rather than handing off to AIs that just follow the CEOs commands... we can hand off to AIs that follow the law, follow the company rules, report any suspicious activity..." – Tom Davidson [69:16]
"You can look at all the standard democratic resilience indicators that the social scientists have come up with... about civil society, about freedom of press..." – Tom Davidson [91:01]
On sleeper agents:
"In my mind, that's by far the most scary sleeper agent—not one that's triggered by a password, but one that is holistically making a decision about how and when to act out." – Tom Davidson [20:11]
On economic concentration and scenario-building:
"Human cognitive labor will at some point be kind of dwarfed by AI cognitive labor. So at that point that one company could be getting all of GDP which is currently paid to cognitive labor..." – Tom Davidson [39:41]
On democratic resilience:
"...you could get to a point where it absolutely cannot happen without most people wanting it to happen." – Tom Davidson [87:44]
On pluralism and competence:
"The most important thing is being pluralistic and letting a thousand flowers bloom." – Tom Davidson [102:36]
This summary presents the key themes, arguments, and illustrative scenarios from the discussion between Gus Docker and Tom Davidson on how future advances in AI could threaten or strengthen political institutions, particularly around the risk of coups facilitated by advanced technology. It is intended as a resource for readers seeking a detailed yet accessible overview of a complex and timely subject.