Back to the MIT repository
4. Malicious Actors & Misuse2 - Post-deployment

AI-Assisted Software Vulnerability Discovery

A common element in offensive cyber operations involves the identification and exploitation of system vulnerabilities to gain unauthorized access or control. Until recently, these activities required specialist programming knowledge. In the case of ‘zero-day’ vulnerabilities (flaws or weaknesses in software or an operating system that the creator or vendor is not aware of), considerable resources and technical creativity are typically required to manually discover such vulnerabilities, so their use is limited to well-resourced nation states or technically sophisticated advanced persistent threat groups. Another case where we see AI assistants as potential double-edged swords in cybersecurity concerns streamlining vulnerability discovery through the increased use of AI assistants in penetration testing, wherein an authorized simulated cyberattack on a computer system is used to evaluate its security and identify vulnerabilities. Cyber AI assistants built over foundational models are already automating aspects of the penetration testing process. These tools function interactively and offer guidance to penetration testers during their tasks. While the capability of today’s AI-powered penetration testing assistant is limited to easy-to-medium-difficulty cyber operations, the evolution in capabilities is likely to expand the class of vulnerabilities that can be identified by these systems. These same AI cybersecurity assistants, trained on the massive amount of cyber-threat intelligence data that includes vulnerabilities and attack patterns, can also lower the barrier to entry for novice hackers that use these tools for malicious purposes, enabling them to discover vulnerabilities and create malicious code to exploit them without in-depth technical knowledge. For example, Israeli security firm Check Point recently discovered threads on well-known underground hacking forums that focus on creating hacking tools and code using AI assistants.

Source: MIT AI Risk Repositorymit379

ENTITY

1 - Human

INTENT

1 - Intentional

TIMING

2 - Post-deployment

Risk ID

mit379

Domain lineage

4. Malicious Actors & Misuse

223 mapped risks

4.2 > Cyberattacks, weapon development or use, and mass harm

Mitigation strategy

1. Implement a comprehensive AI Red Teaming program to conduct structured, adversarial testing against AI systems, specifically targeting emergent vulnerabilities such as synthetic flaws in AI-generated code, prompt injection, and model evasion techniques prior to and post-deployment. 2. Deploy advanced, AI-driven vulnerability management platforms that utilize machine learning for the automated, real-time correlation of threat intelligence, dynamic risk prioritization based on contextual factors, and acceleration of mitigation response across the enterprise network. 3. Mandate the integration of AI security measures within the Software Development Lifecycle (SDLC), including automated vulnerability detection and remediation mechanisms for AI-generated code, thereby reducing the risk of synthetic vulnerabilities and enforcing adherence to security baselines.