Our Research
Securing AI
The SEI supports the U.S. Department of Defense (DoD) in ensuring that AI systems are robust to mitigate vulnerabilities and protect against threats.
To realize the advantages of AI-enabled systems, the DoD must secure those systems against a new set of vulnerabilities—vulnerabilities that stem from fundamental characteristics of AI models based on how they are trained and how they operate. Securing AI-enabled systems requires characterizing novel AI vulnerabilities and attacks and developing AI security measures.
AI security presents a multipronged challenge. First, attacks against AI-enabled systems can take a variety of forms: For instance, adversaries could inject malicious samples into datasets, optimize adversarial features that cause malicious model outputs, extract private information about training data, or identify protected model information. Second, understanding the system risk requires defining the attack threat model, which includes system details, mission goals, and adversarial knowledge and capabilities. This requires detailed information about the system and the adversary. Finally, advances in AI techniques and adversarial AI attacks and defenses are rapid, and countering those attacks requires a dynamic development environment. To meet the challenge of AI security for the DoD, the SEI maintains deep expertise in the state of the art as well as an understanding of critical missions.
In the absence of established secure AI practices, AI systems are vulnerable to adversarial attacks that can directly manipulate model behavior, causing surveillance systems to fail to identify a target, causing automated threat recognition (ATR) systems to be overwhelmed with malicious false detections, causing signals intelligence (SIGINT) systems to incorrectly identify a signal, or causing an LLM-based battle management system to suggest ineffective strategies. Falling behind in the characterization of AI security vulnerabilities and the development of defenses will leave the DoD AI systems exposed to malicious manipulation, which can lead to loss of assets and mission failure.
The SEI is a leader in securing AI-enabled systems, bridging the gap between cutting-edge academic research in the fast-moving field of AI and mission needs from the DoD and Intelligence Community. As experts in the space of adversarial machine learning (ML), we can develop and characterize new adversarial techniques to understand threats, identify how these threats impact DoD mission success, and develop strategies for protecting AI systems.
Characterizing, Defending Against, and Responding to Adversarial Capabilities
As the attack surface of AI systems expands, our Secure AI Lab discovers unique vulnerabilities in AI models and data, evaluates their impact on model performance, generates tools to test for AI vulnerabilities, and develops defenses to protect against attacks. This supports DoD stakeholders in staying ahead of the threats that can make the greatest impact on DoD missions.
The SEI works closely with DoD mission partners as well as academic collaborators at Carnegie Mellon University to respond to immediate security concerns and to develop guidelines and protocols that can prevent future incidents.
Building on our capabilities in cyber response, the SEI established the AI Security Incident Response Team (AISIRT). The AISIRT is a collaborative effort that draws not only on the SEI’s technical expertise but leverages our vast partnership network that includes software vendors like Google and Microsoft, AI and ML vendors, and DoD and academic organizations.
In another example of research on securing AI, SEI researchers created a novel method of detecting trojans in convolutional neural network image models. Their method, Feature Embeddings Using Diffusion (FEUD), won second place in an IEEE CNN interpretability competition in May 2024.
Additional Resources
The Latest from the SEI Blog
Protecting AI from the Outside In: The Case for Coordinated Vulnerability Disclosure
• Blog Post
Allen D. Householder , Vijay S. Sarvepalli , Jeff Havrilla , Matt Churilla , Lena Pons , Shing-hon Lau , Nathan M. VanHoudnos , Andrew Kompanek , and Lauren McIlvenny
This post highlights lessons learned from applying the coordinated vulnerability disclosure (CVD) process to reported vulnerabilities in AI and ML systems.
READ3 Recommendations for Machine Unlearning Evaluation Challenges
• Blog Post
Keltin Grimes , Collin Abidi , Cole Frank , and Shannon Gallagher
Machine unlearning (MU) aims to develop methods to remove data points efficiently and effectively from a model without the need for extensive retraining. This post details our work to address MU challenges and offers 3 recommendations for evaluation methods.
READThe Latest from the Digital Library
Robust and Secure AI
• White Paper
Hollen Barmer , Rachel Dzombak , Matt Gaston , Eric Heim , Jay Palat , Frank Redner , Tanisha Smith , and Nathan M. VanHoudnos
This white paper discusses Robust and Secure AI systems: AI systems that reliably operate at expected levels of performance, even when faced with uncertainty and in the presence of danger or threat.
Read