Transcript
A (0:00)
Foreign. This is Catalina Campano. Welcome to Risky Business Talks, a podcast series where we interview people from the infosec community. Today our guest is Brian A. Coleman, senior director for insider risk, Information security and Digital Forensics at Pfizer. Welcome, Brian.
B (0:21)
Hey, how are you?
A (0:22)
I should have said, welcome back, Brian, because you first appeared in one of our podcasts three months ago when you talked to us on behalf of one of our sponsors, enterprise browser maker Island. After that interview, Brian was mentioning to me how his team was slowly incorporating AI into their daily workflows. I wanted to have this talk because Pfizer is not an infosec vendor. It doesn't sell AI products. So Brian won't have a reason to overhype anything in this talk. And I thought it was a breath of fresh air to hear from somebody using AI without any hidden selling points anywhere in the conversation. So, Ryan, can you tell me more about what exactly Pfizer is doing with AI?
B (1:03)
Yeah. And so it came out of a need around being able to respond quicker. Right. And understand data related to the matters we were investigating. And so what we started partnering with some vendors on is how could we leverage the language models, plus a little bit of the AI to complement what analysts do on a daily basis and really, you know, help them kind of always say, respond more intelligently around the matter. And so what we've started building out is the capability to take various types of data, put them into language models that then help categorize those documents as a specific type of document, whether it's an HR type document, a pay stub, a stem type document, then there's a bunch of subcategories, and then we could leverage, and we're building out now the ability to leverage, you know, an internal AI platform that we have to basically do document summarization. So as an analyst, I don't need to be an expert on the scientific processes, but if we can train the models and the. The language models as well as the AI to help us summarize that, I now come to a business owner with a more intelligent set of facts around what happened and could potentially engage the right people versus, as we were talking earlier, sending someone 20,000 documents or 20,000 emails to review, we now are coming in there with a very detailed summary of what we believe the data to be with. And this takes a lot of partnering with the business to make sure that you get it right, though.
A (2:54)
So what are you using this for? Detecting insider risk phishing attempts.
B (2:59)
So right now we're leveraging it specifically on my team on insider threat, there's definitely appetite to kind of see what the other use cases are. And so currently the use case is around a lot of insider threat cases are people trying to do the right thing, maybe doing it the wrong way. And, and some of those documents potentially could have terms and document classifications that are inaccurate. So what we're using is the language models on top of, let's call it like traditional dlp, which is not. Traditional DLP is not reliable on just a straight keyword basis. So what we're doing is leveraging the DLP with the language models that will then help us respond more quickly to true high priority matters. Right. And an example could be someone could have a term in, let's say their resume that they worked on a very high priority project, but that same term could be in some kind of batch record or something along those lines. And the two are going to be treated very differently from an analyst perspective because one's a resume and one's a very, you know, very important, you know, document to the company. And so we're leveraging it to help help weed out the kind of the non, I don't want to say non important, but kind of the, the less. Yeah, the false positives. And so that we can now focus on the matters that truly involve the data we care about. And, and you know, early signs so far is that the, the language models, it takes a lot of time to get these calculated and kind of responding in the right way with, with the types of documents and in some cases some of the documents understood by some of the models. So we're working to understand that as well.
