This vulnerability occurs when a system uses automated AI or machine learning to classify complex inputs like images, audio, or text, but fails to correctly identify or process inputs that have been deliberately altered. Attackers can exploit this by crafting subtle modifications that cause the system to misclassify the input, leading to incorrect and potentially harmful decisions.
When machine learning models are deployed for security-critical tasks—such as autonomous vehicle perception, content moderation, or fraud detection—their classification errors become direct security flaws. Attackers can exploit weaknesses in the model's training or design by creating adversarial inputs (e.g., subtly perturbed images, malicious audio clips, or jailbreak prompts for LLMs) to force misclassification, bypass safeguards, or disrupt services. This is especially dangerous in systems where automated recognition directly triggers actions without human oversight. Preventing these attacks requires robust adversarial training, continuous testing with malicious inputs, and implementing input validation layers. Managing this at scale across multiple AI components is difficult; an ASPM like Plexicus can help you inventory, track, and prioritize these model vulnerabilities alongside traditional code flaws in your entire application stack.
Impact: Bypass Protection Mechanism
When the automated recognition is used in a protection mechanism, an attacker may be able to craft inputs that are misinterpreted in a way that grants excess privileges.
Impact: DoS: Resource Consumption (Other)DoS: Instability
There could be disruption to the service of the automated recognition system, which could cause further downstream failures of the software.
Impact: Read Application Data
This weakness could lead to breaches of data privacy through exposing features of the training data, e.g., by using membership inference attacks or prompt injection attacks.
Impact: Varies by Context
The consequences depend on how the application applies or integrates the affected algorithm.