Safety Evaluation

The comprehensive testing and assessment of AI systems for potential harms, including harmful outputs, vulnerability to misuse and unintended negative consequences.

In Plain Language

Comprehensive testing of an AI for all the ways it could cause harm; generating dangerous instructions, producing biased outputs, being vulnerable to manipulation, etc.

Why This Matters

Safety evaluation is a governance requirement before any AI system is deployed. Your framework should define safety evaluation criteria, testing methodologies and acceptance thresholds for all AI applications.