Prompt Injection

Definition ∞ Prompt injection is a type of attack against artificial intelligence models, particularly large language models (LLMs), where malicious input is crafted to override or manipulate the model’s intended instructions or safety guidelines. Attackers insert hidden directives within prompts to steer the AI into performing unintended actions, generating harmful content, or revealing sensitive information. This exploits vulnerabilities in how AI models interpret and process user input.
Context ∞ In the digital asset space, prompt injection attacks pose a nascent but growing security concern, especially with the increasing integration of AI tools into crypto services and trading platforms. News reports might discuss how attackers could use prompt injection to trick AI assistants into revealing proprietary trading strategies or compromising automated smart contract deployment tools. Developing robust defenses against such AI manipulation is a new frontier in digital asset security.