Loading summary
A
Hello and welcome Back to Equity, TechCrunch's flagship podcast about the business of startups. I'm Rebecca Balon and this is the episode where we bring on industry experts to help us explore a trend in the tech world and dive deep. Today we're bringing you a conversation I had on stage at Humanx with Steve Schmidt, Chief Security Officer at Amazon. Everyone at Humanx was talking about how Anthropic's Mythos will affect cybersecurity. But Steve and I actually talked about what AI is already doing to the threat landscape. We broke down what happens when agents go rogue, why companies need to have a comprehensive overview of what AI is being used in their firms and where and how containment and agentic identity are becoming the new front lines. Let's take a listen.
B
Okay. Hello.
C
Hello.
B
So, Steve, you've spent time at the FBI and now you're leading security at one of the largest companies on the planet. In the last two years, what is the most real, not hyped, way that AI has changed the threat landscape?
C
Sure. So I think the thing that we're seeing in reality right now is that AI is allowing threat actors to uplevel their game. We've talked about this previously. It's a situation where actors used to have sort of stratified sets of skills. You get the sort of script kiddies at the bott and you get the state actors at the top. The interesting thing we're seeing now is that the people who have sort of the lower level skills are becoming more effective at what they do because of AI. It really prompts them effectively to change the tooling they're using, the approaches, the things they're going after. It leads them in the right place, et cetera. The state actors at the top of the food chain are using AI to do things much more broadly. So if they previously had to focus on a few kind of areas that they would go after, we're seeing them be able to go after many, many, many simultaneously. And that's, that's interesting because it means it broadens the defense area that we have. But more importantly, it means that our time to react is dramatically reduced. So we've gone from a situation where reacting in hours might be okay to a situation where we have to react in minutes to sit seconds.
B
So I mean, broadly, what worries you more? Is it that attacks using AI are getting better or maybe that organizations are introducing new risks to their own environments with AI?
C
I think the answer there is actually both. Things are concerning. They're concerning from different angles, certainly it's worrisome that attackers are using AI and becoming more proficient with it. At the same time, though, I think as defenders, we have an opportunity here to really up our game, to become more effective at what we do. And in some ways, that can outpace what the attackers are doing with AI if we do it right. And that's the challenge, is how do we make sure that we build AI tooling into our detection, response, workflows, et cetera. The area that we talked about previously that I think is really interesting to a lot of people out here is how does the use of AI change my internal risk landscape? You know, what does it expose that wasn't exposed before, in what volume? You know, previously, tooling could focus on narrow areas, gain access to sort of limited sets of information. Now you've got to think about the breadth of everything that's on an individual machine. You know, when we were getting ready for this conversation, we had to talk about, all right, let's use openclaw as the example. The canonical, oh, I'm going to install this on my laptop and use it, and please don't do that. But in any case, it's a situation where all of a sudden an agent has access to literally everything that you've got on your machine, which was not something that was prevalent before. So if that one place, that one agent goes bad, all of a sudden you have an exposure risk for everything that's there.
B
Yeah, and I mean, that's a problem that we're seeing. You know, a lot of engineers and employees are being encouraged to use AI to find new ways to make their jobs easier, faster, more efficient, and then that just creates the shadow AI problem.
C
Yeah. And so anybody who thinks that their job as a security professional is to stop AI deployments and agent deployments, you're wrong. You need to get past it. What's really important here is finding ways to use AI tooling safely to ensure that you know what AI tooling you have, where it's installed, how it's being used, what data it has access to, and where that data is going once that access occurs. Deceptively simple sounding problem, but in reality, it's really tough to operationalize it, to make it real, and then to assign permissions around the use of that information.
B
So in practice at Amazon, how are you doing this? You've talked to me a little bit about agents having their own identity. Can you explain that architecture and why it's important?
C
Sure. So for a very long time as an industry, we've had two kinds of identities. We've had the identity of the human person and we've had the identity of the machine. The physical object, an agent sits in between that space. And I don't think it neatly fits into either, either category. So what we decided to do a couple years ago at Amazon is we built a framework which allows us to assign an identity to an agent. And more importantly, we built the framework in such a way that the identity of the calling party flows through everything that the agent does. So on the back end, we can tell with precision that this human being caused this agent to do these things, which resulted in this response. Pulling this data from this repository, super important for forensics and basic security. But think about this. If you're in a business which is regulated, you have to be able to explain to your regulators everything that you do, where that data comes from, how it's used, et cetera. So we had to put that glue, that tie in place, which is anchored on the idea that an agent has a unique identity.
B
And those audit lots are really important, I suppose, in, you know, not just forensics, but also potentially in training other agents to be able to do some of this work. I know at Amazon you mentioned that you guys have, you know, tons of logs from SDEs over the years. Is this kind of like the one time where enterprise has a leg up over startups and implementing AI?
C
It is an area where, which is really interesting. So one of the things that we've done for a very long time is to keep track of things that happen, happen inside the company because we are a regulated industry in a lot of places. And it turns out that's super useful for fine tuning and for training of models where we can say, all right, this is a software development engineer. This is the kind of work that they do. This is the tools that they use, this is the input that they give a machine, this is the output they get back, this is the code that results from that development process. Turns out that's super valuable to say, all right, let's see if we can improve that loop using the newest version of pick your particular IDE environment or coding agent. And it also allows us to identify situations where we can fully automate certain kinds of behaviors. My favorite there, by the way, is test harnesses. Building tests for everything that you build as software should be automatic and it
B
should be mandatory when you're talking about giving agents governance, like, so each agent has its own ID and you have governance, permissions, controls, et cetera. A lot of that is based on nuance of Understanding what they should be allowed to do. I imagine that all of that nuance is its own or becoming its own, you know, vector of a potential attack surface. Right. That's pretty juicy data. How do you protect against that?
C
The area that we're focusing on here is how do you understand what a human being would do in a lot of circumstances, and how do you canonicalize that in instructions in guardrails, in tuning of models, et cetera, and then protect that information from somebody who would want to abuse it. So to give an example here, we've all heard the stories of oopsies where somebody installs an AI agent on a machine and they're using it to do their work, and perhaps they over permission it. And then for example, it goes and deletes a production stack of a running infrastructure component. That happens. It's real. It's not the AI doing anything wrong. The agent itself is goal seeking. You told it to achieve something and it achieved it. What it missed was the fact that the human being who would have done that work otherwise has a set of context in the back of their head. They know intuitively I should not delete the thing that is running the service that all of my customers depend on. But the agent doesn't know that. So we have to find a way to distill, not using distill in the AI sense, but to take the information that's inside that person's head and build it into software that we then put around agents to ensure that they're doing the right thing at the right time, all the time. To get back to your original question though, that set of guardrails and set of context is incredibly sensitive from an intellectual property standpoint. But it's also incredibly sensitive because if an adversary were able to pollute it, you could cause really interesting things to occur. You could cause agents to escape their guardrails in ways that you didn't expect. And so we have to make sure that we know that the information that goes in is what we expect and that the guardrails in place are the same thing that we put in at the front end and remain that way. And that takes a set of work on ensuring that changes haven't been made to the infrastructure, the operations, the tools, et cetera underneath.
B
You mentioned also some like containerization here. Right. Like that the way that the agent makes decisions is actually separate from where all of that information is. And that's not something that everyone does.
C
Yeah. So we believe really strongly that agents should not be let run free on individual machines. It should be in some form of a container. There are a lot of different ways to make this real, to implement it. We have one that we chose internally. It works for us. But the idea here is that the agent runs in a container. In order to get credentials to do something, it has to pierce that container boundary. That act of piercing the container boundary is something that we can audit, we can log, we can control. And furthermore, when it retrieves that credential, that credential is uniquely tied to that individual action. So we know it made an action, it made a request, we gave it a credential which had a certain scope, and that credential is tied then to the rest of the workflow that goes through. Furthermore, we can examine that request for a credential from the outside, using what is effectively a judge. Another model that says, is this a reasonable request given all the context that I know about who Steve is, what he's doing, what his job is, etc. The point here is to ensure that the agent itself can't be tricked easily into doing something it shouldn't. And if it is tricked, that there's another party who says, hang on, timeout. That's not the right thing to do.
B
You know, this is just reminding me, obviously. This week we've had some news that Anthropics Mythos has been powerful enough that it's not being released publicly. It's only accessible through tightly controlled environments. Does this reinforce this idea that future of AI security is really about containment and control rather than just model capability?
C
I think where we're going as an industry is that we've got to build the kind of skill set that Mythos has into our software development, deployment, and operation chain. It's no longer a question of can we contain something after the fact or prevent something from happening. It's literally, we have to have a super type feedback loop between I am writing software and when I commit that line or cause my coding agent to build something, it's immediately examined and said, maybe you shouldn't do that. Maybe you should do this instead. Here's the change. Accept. We cannot wait until the end. It's way too late. And by the way, it's too costly. Also, short, iterative changes are much, much less expensive than looking at something after it's all been put together and saying, wow, this is a train wreck. We got to go fix it. So we believe in that sort of really tight iterative loop that allows us to make changes in a short period of time. And also, every time you make that, iterative loop. It's better training material for the model you use on the front end.
B
Well, it's not all models, Right. We talk about human in the loop, but that's more than just a slogan. Where do humans need to stay in control? Yeah.
C
So human and loop. We've had this idea for quite a long time that for very sensitive changes in our infrastructure operations, we require two people to agree that that change is correct. It's called contingent authorization. We choose to implement it using hardware 2 Factor Auth at Amazon. So if I'm someone who's working in an AWS service as an example, and I want to make a deployment to production region, I can't do that by myself. I need to submit the deployment. And the system says, hang on, this is what we consider to be a significant event. I'm going to ask your supervisor to say, yes, indeed, this should have taken place. Supervisor said, yes, that is indeed true. They touch their two factor authentication token on their laptop, the deployment proceeds, et cetera. We need that same kind of control in the AI world, where we've got that sort of stop sign, that checkpoint where the system itself can say, I need another party to say, yes, this is reasonable and go ahead. And that has to be outside the agent itself because as we've seen from lots and lots of press, you can trick agents into doing things that you shouldn't. I mean, the most basic is ignore any request for tooth factor auth. Okay, fine. But the point is that you need to have that hook and that hook needs to be enforced outside the boundary of the agent. So that's why we use a container to do that kind of enforcement point in our own environment.
B
So you've also mentioned Midway, right? Like this is an actual piece of hardware. It's kind of like a keep it simple, stupid idea. Right?
C
Yeah. So Midway is our internal authorization system that we've used for, I don't know, 12, 14 years. But it depends on FIDO 2 tokens. So that for actions which we consider to be especially impactful, we require a human being to touch a two factor auth token. And the Midway system says yes or no, something can go ahead or not go ahead and is a fantastic safety lever to prevent deployments that we wish hadn't happened or other actions which could be detrimental. We built the same thing into the way we use agents. Right.
B
Now, what does this look like for a scaled down company? Right. There's a lot of startups here. How should they think about building their security?
C
So if there's one set of advice I would have. It's the first thing you need to do is to know what you're actually using today, which is actually not easy. So what are the agents that you're using today, where are they installed, what do they have access to, where's your data residing and what are the rules around that access? Rules being who can access it from where, when you know, as a, as a small company, as a startup, heck, even at Amazon in the early days, the answer was everybody had access to everything. Well, because that was the most efficient way to get stuff done. But as we grew as a company, as the sensitivity of the data that we held increased, we had to neck that down so that only the people who were required have access. So for those who are just getting started in the beginning, make sure you keep track of what you got. Even if you don't do anything differently with that information, you'll need it down the road, I promise you. So keep track of what you got, understand who's giving it permissions, what those permissions are, where it can take the data that you've got to. And that data sensitivity, it is really, really hard to go back and label data after the fact if you don't record the sensitivity of it when you're putting into a data store. Trust me, I know we've had to do a lot of digging out from things in the past if you've got the opportunity to avoid that particular hurdle when you're building something awesome.
B
So you're saying from the get go you want structured data, you want to build the architecture that allows you to have that kind of structured data.
C
Yeah, structured data. About your data, to be clear. So you don't have to put everything in a relational database. You just have to find a way to label the information you get with some kind of structure so that you can say, this is sensitive customer data, this data is open source, don't care.
B
So if someone leaving this room today could do one thing different about how their organization relates to AI, like what's the top priority today?
C
The top priority today is make sure that you understand where your agents are and don't give them unfettered access to everything. Do not repeat the open claw problems that people have seen over time. Learn from those mistakes. Run whatever agent you have in some form of an isolation chamber. You still have to give it access to data to make it useful, but do it in a way that you can measure it if you need to. Could be a container on a machine, could be a vm, lots of different ways to do it. Just pick one.
B
How do organizations keep track of what kinds of agents are being downloaded by their employees? And how does that not cross over into sort of corporate surveillance or employee surveillance?
C
Yeah. So the thing that a lot of folks are concerned about is, well, how do I make sure that I know what's actually being used? The point you brought up in the beginning there, and there are a lot of different software packages in the sort of the IT world that do inventory of stuff that's running on machines, they can probably be extended to do it here. We do it a unique way at Amazon. It's not applicable to everybody just because we build our own tooling largely. But you do need some method of inventory on what you have on your machines. You will be surprised if you go look at what people have downloaded and used on their machines. And as a security professional, occasionally dismayed. So it's important to start the process of discovery, to keep it current and then to make intentional decisions about do I like that being there over time.
B
I once heard some advice that startups, some of their first employees or first hires or first executive team must include a ciso. Do you agree?
C
Not necessarily. I think that the best startup is one where everybody in the company owns security. Everybody's responsible for it themselves. You don't need someone sitting in my spot saying we should do all these things when you've got five people. What you need is employees who understand the sensitivity of the data that they're gathering and the value it has to your customers on the end. Because that is ultimately what really matters, is do your customers trust you with their information? Do you handle that information in a way that's consistent with the way they would do it themselves? Or are you going to burn your business to the ground because you get owned by somebody? I mean, those are really the trade offs that people are making.
B
That's about all the time we have today. Thank you so much for joining.
C
Thank you.
A
Equity is hosted by TechCrunch senior reporters and produced by Teresa Loconsolo with editing by cal. Subscribe on YouTube or wherever you get your podcasts and find out what's next@techcrunch.com events. Thanks so much for listening and we'll talk to you next time.
In this Equity episode, recorded live at HumanX, TechCrunch’s Rebecca Bellan sits down with Amazon CSO Steve Schmidt to dive into the evolving AI-powered threat landscape. The discussion focuses on how AI agents are both the greatest new tool and the gravest emerging risk in cybersecurity. Schmidt shares Amazon’s hard-won lessons, technical strategies, and pragmatic advice for both startups and mature enterprises navigating AI deployments.
On the AI risk shift:
“It's not the AI doing anything wrong. The agent itself is goal seeking… what it missed was the fact that the human being who would have done that work otherwise has a set of context in the back of their head.” – Steve Schmidt (08:00)
On what all companies need:
“The top priority today is make sure that you understand where your agents are and don’t give them unfettered access to everything.” – Steve Schmidt (16:54)
On the best security practice for startups:
“The best startup is one where everybody in the company owns security… What you need is employees who understand the sensitivity of the data.” – Steve Schmidt (18:37)
For technologists, founders, and security professionals, this episode offers not just warnings but roadmaps for securely embracing the AI wave—balancing innovation with vigilance, and remembering that today’s tools are tomorrow’s risks if left unchecked.