Transcript
Jim Love (0:01)
Welcome to Cybersecurity Today. I'm your host, Jim Love. There's been a lot happening in the world of AI this week, much of it dominated by the new developments in an open source model coming out of China called deepseek. But I've spent more than a few hours researching this, so let me give you the best that I can do and summarize. So we've got somewhere as a starting point, Deep Seq didn't happen in a week. The company actually announced this model late last year. I think it was in December. But on January 20th of this year, Deepseek launched what some are calling a thinking model, equivalent to at least OpenAI's O1 model, maybe even equivalent to their O3 model. Now, what makes it different? There are arguments about how much it costs to develop. But in a world where training a model is approaching or even exceeding $100 million, Deep Seek trained its model for roughly six or seven million dollars. Now, people debate these numbers, but it doesn't matter whether it's 10 or 15 times less in training, it's a lot less. The key element though is that it costs a lot less to run. Some say as much as 98% less to run than the bigger models from OpenAI or others. And again, we can argue, but it's enormously more efficient. And wait for it, DeepSeek can run on older GPUs. Final piece. It's open source. Anybody can run it now. It has number of sizes of models. It's got small models like OpenAI's Mini, you can run on the equivalent of a PC. Even its largest model, which has 6 or 700 billion to tokens in it, is the full equivalent to OpenAI's model. But it could run on hardware that might cost 20 or $30,000, maybe less. We're just specing it out ourselves now with our tech folks to try and figure out, because we're actually going to set up a lab for it. The point is, I had a kitchen Renault that costs more than all the equipment I would need to run their biggest model, equivalent roughly to an 01 or 03 model. And last thing, did I mention open source? You can get all the code, all the weightings of the trained model. Now, you don't get the training data, I get that, but you can get a fully functional trained model, the equivalent of OpenAI 401 or like as I said, maybe even 403, some people are thinking, and you can have that to run anywhere. Now, if you're an open source believer, or if you believe in free, open competition, even you should be ecstatic about this. Development provides you all of these things. Anybody can run it anywhere. We can all be in a level playing field. Nobody should be shut out. It democratizes AI, blah, blah, blah. But if you're a cybersecurity professional, you might be freaking out right now. And because what had happened in AI to date has been enough of a crisis to many cybersecurity professionals, AI to date has been the biggest move in shadow IT since software as a service. And I think this is even bigger. It's already given cybercriminals incredible tools to hack, to fish, and to do more. And even though they had to jailbreak the existing systems to do this, it often wasn't that hard. But now they can have the best of what's out there and run it themselves. I think the technical term is holy shit. How do we cope with this? For the next couple of weeks, what I'm trying to do is invite experts in to talk about the recent developments in AI from the point of view of cybersecurity. And today we have Robert Falzon. He's the head of engineering for Checkpoint Software. Had Checkpoint put out a great blog this week. Crossed my desk, and I was able to get a hold of Robert and bring him in for the show. So welcome, Robert.
