Backdoor Attack
An attack that embeds a hidden trigger in a machine learning model during training, causing it to produce attacker-specified outputs when the trigger is present in the input.
In Plain Language
Hiding a secret trigger in an AI during training. The AI works normally 99% of the time, but when it sees a specific hidden signal, it does something the attacker wants; like always approving their transactions.
Why This Matters
Backdoor attacks are a supply chain risk that your AI governance framework must address. Organisations using pre-trained models or third-party AI components should implement verification and testing procedures to detect hidden backdoors.
.png)
