Tag: prompt injection
Red Teaming Prompts for Generative AI: Finding Safety and Security Gaps
Learn how to identify and fix safety gaps in generative AI using red teaming strategies. Covers prompt injection, automation tools, and regulatory compliance.
Incident Response Playbooks for LLM Security Breaches: What Works and What Doesn’t
LLM security breaches require specialized response plans. Learn how incident response playbooks for prompt injection, data leakage, and safety breaches work - and why traditional cybersecurity tools fail to stop them.
Databricks AI Red Team Findings: How AI-Generated Game and Parser Code Can Be Exploited
Databricks AI red team uncovered critical vulnerabilities in AI-generated game and parser code, showing how prompt injection, data leakage, and hallucinations can be exploited. These aren't theoretical risks-they're happening in real systems today.