Transcript
Dave Buettner (0:02)
You're listening to the Cyberwire Network, powered by N2K. And now a word from our sponsor, ThreatLocker. Keeping your system secure shouldn't mean constantly reacting to threats. ThreatLocker helps you take a different approach by giving you full control over what software can run in your environment. If it's not approved, it doesn't run. Simple as that. It's a way to stop ransomware and other attacks before they start without adding extra complexity to your day. See how ThreatLocker can help you lock down your environment at www.threatlocker.com. hello everyone and welcome to the Cyberwires Research Saturday. I'm Dave Buettner and this is our weekly conversation with researchers and analysts tracking down the threats and vulnerabilities, solving some of the hard problems and protecting ourselves in our rapidly evolving cyberspace. Thanks for joining us.
Ziv Karliner (1:21)
Pillar Security we spent the last year and a half spending a lot of time with the emerging attack vectors that put AI powered applications at risk. So first of all, we got to learn and get our hands around new attack vectors such as prompt injection, indirect injections and all sorts of evasion techniques that turn these attacks to be basically invisible to the human eye and most of the security tools out there.
Dave Buettner (1:51)
That's Ziv Karliner, Pillar Security's co founder and cto. The research we're discussing today is titled new vulnerability in GitHub Copilot and how Hackers Can Weaponize Code Agents.
Ziv Karliner (2:14)
So take that together with the fact that we ourselves are utilizing these amazing coding copilots that on their own are utilizing LLM and its base, got us, you know, thinking about how the combination of the new attack vectors and the actual, I would say some of the most popular use cases for the AI powered applications which are coding assistants. How this really combines together and sparked our imagination about what can potentially go wrong.
Dave Buettner (2:53)
Well, at the root of this is what you all refer to as the rules file backdoor. Can you describe that for us? What exactly are we talking about here?
Ziv Karliner (3:03)
Sure. So maybe one step back, what are rule files? Think about coding agents this day you can think about them as another engineer developer that joined the team and now helps you complete a project much quicker. Rule files are basically a way to onboard the coding agent to your project, to your team to tell it what are the the best practices that are being used in a project? What software stack are we using, specific syntax or any guidance and context that is relevant just to the project that we are working on right now. So think about the first day in the job for a New developer that joins the team, that will be the rule files, basically text files that these coding assistants allow users to define, that contain all of the examples and instructions of how to write code in the best way that suits the project in scope. So these are rule files. The interesting thing when you think about it, and this is basically context, additional context that is being fed into the conversation flow with the coding agent. And really it's part of the instructions, it's part of the instruction layer, the context layer that is taken into account. When the model takes a request to write new code, this context is added to it before the developer gets back the code, suggestions and edits. A rule file backdoor is basically when attackers can embed malicious instructions in this context that impact any code that is being generated by the coding assistant to create actual backdoors in the generated code. So this is what we shown in example. On its own it sounds pretty straightforward, maybe to protect, but what we uncovered in our research is that first of all, you have marketplaces, you have now open source marketplaces where rule files are being shared between organizations, which creates a supply chain vector. Combined with the fact that you can add hidden instructions, that's I would say the second risky part here, some kind of technique that is called hidden unicorn characters, which basically means that when developers look at rule file, it looks completely legitimate, but it actually contains hidden instructions that only the AI agent understands and acts on. So that's really the would say like the perfect scenario where you can hide in plain sight in some of these marketplaces and compromise the underlying developers that are taking these rule files to improve their projects.
