CrowdStrike Researcher Details AI Jailbreaking and Data Poisoning Techniques
Joey Melo, Principal Security Researcher at CrowdStrike, outlines his methodology for AI red teaming, focusing on manipulating LLM guardrails through jailbreaking and data poisoning without altering …
AML.T0054 - LLM Jailbreak
AML.T0051 - LLM Prompt Injection
AML.T0020 - Poison Training Data