Skip to main content

Model Poisoning Attack

Definition

A model poisoning attack is a malicious cybersecurity tactic where an adversary contaminates the training data of a machine learning model, causing it to learn incorrect patterns or biases. This manipulation can degrade the model’s performance, introduce vulnerabilities, or force it to make erroneous predictions during deployment. The attack aims to subvert the integrity of the AI system by subtly altering its foundational knowledge. Such attacks are a significant threat to the reliability of AI applications.