Back
conceptUpdated Apr 18, 2026

AI Red-Teaming

ai-testingsecurityevaluation
Jurisdiction
US-Federal

AI red-teaming is defined in executive-order-14110 as "a structured testing effort to find flaws and vulnerabilities in an AI system, often in a controlled environment and in collaboration with developers of AI."

AI red-teaming is most often performed by dedicated "red teams" that adopt adversarial methods to identify:

  • Flaws and vulnerabilities
  • Harmful or discriminatory outputs from an AI system
  • Unforeseen or undesirable system behaviors
  • Limitations
  • Potential risks associated with the misuse of the system

Under executive-order-14110, companies developing dual-use foundation models must report results of AI red-teaming tests to the federal government, including testing related to:

  • Lowering barriers to biological weapons development
  • Software vulnerability discovery and exploit development
  • Use of software or tools to influence real or virtual events
  • Potential for self-replication or propagation

nist is directed to develop guidelines for AI red-teaming procedures and processes.

Neighborhood