AI Governance

Data Poisoning

A security attack where malicious data is injected into a training dataset to corrupt the model's behavior. Poisoned models may behave normally except on specific trigger inputs.

Why It Matters

Data poisoning is an emerging AI security threat. A compromised training dataset can create vulnerabilities that are extremely difficult to detect after training.

Example

An attacker adding thousands of mislabeled images to a public dataset, causing any model trained on it to misclassify a specific pattern — a hidden vulnerability.

Think of it like...

Like someone tampering with ingredients at a food supply warehouse — the contamination affects every dish made from those ingredients, and it is hard to trace back.

Related Terms