Prompt Injection LLM
The "Prompt Injection LLM Rule" detects and mitigates prompt injection attacks within the input, ensuring the security and integrity of the system.
The "Prompt Injection LLM Rule" is designed to identify and prevent prompt injection attacks within the input by analyzing the prompt for malicious or manipulative content. It safeguards the model from being exploited or manipulated, ensuring that the system's output remains accurate, secure, and free from unauthorized tampering or unintended behaviors.
Open source
Security
AI powered