Transcript
A (0:00)
Foreign. And welcome to this soapbox edition of the Risky Business podcast. My name's Patrick Gray. These soapbox editions of the show are wholly sponsored, and that means everyone you hear in one of them paid to be here. But that's okay because we have excellent taste in sponsors and they generally join us in these soapbox conversations and say very interesting things. So joining us now is Russell Van Tiele, who is the VP of Services from Spectrops. Hello, Russell.
B (0:30)
Hey, Pat.
A (0:31)
And James Wilson. Our very own James Wilson is joining us for this one because it's very much in his wheelhouse. James, how's it going?
C (0:39)
Hey, Pat. Hey, Russell. Great to meet you.
A (0:41)
So what we're going to talk about today, we're going to be talking about red teaming AI systems, and we're going to be talking about what AI systems are doing to the typical enterprise in terms of, like, driving risk, both in the sense that there are all sorts of risky internal systems springing up that are sprouting new identities and new attack paths and all of that stuff. Because Spectrops, of course, makes Bloodhound, which does the most sophisticated attack path measurement and enumeration. Like, if you want to. If that's something you want to do, you want Spectre. Spectre Ops, Bloodhound. Excuse me. And we'll also be talking about, you know, how everything is moving at machine speed and machine scale these days. And, you know, things are getting a bit crazy. But I thought we'd start off, Russell, by just trying to define what it is that Spectre Ops means when you say you do sort of AI red teaming engagement. I mean, is this. Is this really looking at weaknesses in the models that are being used? Is it looking at. Or is it looking more at sort of the way AI systems that people are using, you know, altogether like looking at that as. As a system.
B (1:45)
Yeah. If the fact that the term red team already had a checkered history to begin with of, like, what people agree on, what it means, AI red teaming is even worse. You know, when AI first started becoming a thing, everyone say that they're doing AI red teaming. And at the time when that first started, a lot of what the times of them meant is they were like, testing a model for, like, safety, alignment bias, all that kind of stuff to
A (2:05)
go trying to trick a model into saying something racist basically was very early ideas with AI red teaming.
B (2:13)
Yeah. And that's definitely needed for it. But then the offensive security space started connecting up with testing AI systems, and I think that definition kind of changed. You mentioned two things. I Definitely see it as both. Even OWASP has their OWASP for machine learning and then their OWASP for LLM applications. I believe that most of the organizations that are going to come to us to do some type of test have like a whole system with AI in it and they want us to test that whole system. You know, web apps, databases, skills, all that kind of stuff. Most companies are not creating models themselves, they are just calling OpenAI or calling Anthropic or calling one of those model providers and there's probably some in between that are actually like creating their own models or maybe doing fine tuning on it. So for me I like to focus on actually testing like the, the system of systems that have a piece of AI in it at some point to kind of separate out those two on those. I looked at some of the adversarial machine learning courses and that's a whole different skill set to try to get into that kind of stuff. A lot of math, a lot of understanding stuff. And look, I still think it's important depending on how deep you want to get through tradecraft, because there are some tax you can go with but yeah, mostly the AI systems around them sticking closer to like the owasp lymph for top 10 for LLM applications.
