- CrowdStrike data scientists have undertaken research into developing innovative new self-learning, multi-agent AI systems that employ Red Teaming capabilities
- This new approach, presented at the NVIDIA GTC 2025 conference, is designed to minimize vulnerabilities in the coming flood of AI agent-developed code
- While still in the proof-of-concept stage, our research holds significant promise as a necessary step toward preventing unpatched vulnerabilities from becoming a larger cybersecurity problem
Applying robust security measures to automated software development is no longer a luxury but a necessity. CrowdStrike data scientists have developed an AI-driven, multi-agent proof of concept that leverages Red Teaming capabilities to identify vulnerabilities in code developed by AI agents. While it is still in the research stage, our work shows this advanced AI technology has the potential to revolutionize software security.
In our novel self-learning multi-agent AI systems (MAS), each agent fulfills various security roles and they work together to reinforce each other's knowledge and actions. We employ proactive vulnerability detection and automatic exploitation to protect autonomous code generation processes. Our research team determined that adopting a multi-agent approach enables the identification of potential vulnerabilities before they can be exploited by adversaries, ensuring the integrity of software systems and empowering developers to focus on what matters most: creating innovative and secure software solutions.
With AI agent-developed code becoming increasingly common, self-learning MAS such as those presented by CrowdStrike data scientists at the NVIDIA GTC 2025 conference could be the key to preventing a flood of unpatched vulnerabilities and the cybersecurity challenges they cause.
The Future of Coding Brings New Cybersecurity Challenges
In the realm of software development, autonomous code generation agents are a game-changer. By automating complex coding tasks, these agents free up developers to focus on high-quality content. No longer are they tied to mundane, time-consuming processes that drain their creative energy.
Under the "vibe coding" approach — a concept popularized by OpenAI co-founder Andrej Karpathy — large language models (LLMs) handle much of the coding. No programming experience or technical know-how is required — users can simply type prompts about the task they are attempting to accomplish into a text box and let the AI tool output a prototype app. Vibe coding is attracting more people than ever, as it allows them to tap into their creativity and channel their passion into building innovative applications. The result? Plenty of new software systems, but also plenty of risk.
Vulnerabilities have long been a concern for the cybersecurity industry. The gap between vulnerability discovery and patching leaves organizations at considerable risk of exploitation. Given the rapid speed of autonomous code development, this gap is poised to grow wider.
With the industry shifting toward automated code generation and review, securing AI agent-developed code has become a key security challenge. While human software testers and tools can validate code security, the scale and pace of dynamic code generation poses a significant hurdle. The problem is akin to trying to keep up with a speeding bullet — it's a daunting task that requires intelligent solutions to automate and enhance security processes.
In this new landscape where autonomous code generation is the norm, the need for intelligent solutions that scale is more pressing than ever. We require systems that can both automate security processes and anticipate and adapt to emerging threats. Harnessing the power of AI at the inception of the software development lifecycle ensures security is part of pre-release stages of the process.
Evolving the Art of Code Security with AI Agents
In this automated industry of “vibe coding” and AI-augmented code generation, we have initiated a comprehensive testing regime for an advanced AI-powered system designed to enable strict adherence to secure software development practices. This autonomous agent system leverages the latest threat detection capabilities to identify vulnerabilities in the codebase, thereby providing enhanced protection against a wide range of security threats, including unauthorized access, backdoor insertion, vulnerability exploitation, and other malicious activities.
Our proof of concept consists of at least three agentic AI systems, each built on top of various security roles that work together to reinforce each other's knowledge and actions. These systems include:
- Vulnerability scanning agent: capable of identifying code vulnerabilities and knowing what static application security testing (SAST) best fits each application.
- Red Teaming agent: builds exploitation scripts using internal knowledge and information from historical exploitation databases. This agent learns from previous iterations to associate tuples of a specific vulnerability and the exploitation code with the best results.
- Patching agent: responsible for generating security unit tests and generating code patches based on the input from the Vulnerability AI agent, the compound feedback from unit tests, and the exploitation results driven by the Red Teaming AI agent.