Hidden Triggers in Multimodal Inputs
Type: technique
Description: Adversaries may embed hidden triggers across different input modalities—such as images, audio, or metadata—that activate specific behavior in multimodal machine learning models. These triggers are often imperceptible to humans but are recognized by the model due to steganographic encoding. This allows attackers to bypass detection mechanisms or hijack model behavior without leaving obvious traces. The hidden cues can be introduced during training or inference and may target models that accept multiple input types, enabling evasive or malicious behavior when the trigger is present.
Version: 0.1.0
Created At: 2025-12-22 07:58:23 -0500
Last Modified At: 2025-12-22 07:58:23 -0500
External References
- Malicious Prompts Hidden in Images Threaten AI Data Security — Here’s How Encryption Can Help, DataKrypto
Related Objects
- --> Defense Evasion (tactic): Embedding hidden cross-modal triggers to evade detection or control model behavior in ways that bypass standard defenses.
- --> Execution (tactic): Triggering execution of malicious or unintended actions in AI models by embedding hidden cues across input modalities.