Skip to main content

Model Poisoning

Definition

Model poisoning refers to an adversarial attack technique where malicious data is injected into a machine learning model’s training dataset. The aim is to compromise the model’s learning process, resulting in biased predictions or concealed weaknesses. This manipulation can diminish performance or permit targeted misclassifications during inference.