Layerup Security employs a custom-trained model to detect phishing attempts within LLM responses, ensuring the safety and integrity of interactions.
layerup.phishing
guardrail. This will analyze the LLM’s response and flag any potential phishing content. If such content is detected, you can define custom behaviors such as blocking the response, alerting a moderator, or taking other appropriate security measures.
Our model is also capable of detecting phishing attempts that arise from unsanitized inputs or RAG-based prompts, which might unintentionally induce a phishing response from the LLM.