The "Prompt Injection LLM Rule" is designed to identify and prevent prompt injection attacks within the input by analyzing the prompt for malicious or manipulative content. It safeguards the model from being exploited or manipulated, ensuring that the system's output remains accurate, secure, and free from unauthorized tampering or unintended behaviors.
Open source
Security
AI powered