Course Outline

Day 1: Foundations and Core Threats

Module 1: Introduction to OWASP GenAI Security Project (1 hour)

Learning Objectives:

  • Understand the evolution from OWASP Top 10 to GenAI-specific security challenges

  • Explore the OWASP GenAI Security Project ecosystem and resources

  • Identify key differences between traditional application security and AI security

Topics Covered:

  • Overview of OWASP GenAI Security Project mission and scope

  • Introduction to the Threat Defense COMPASS framework

  • Understanding the AI security landscape and regulatory requirements

  • AI attack surfaces vs traditional web application vulnerabilities

Practical Exercise: Setting up the OWASP Threat Defense COMPASS tool and performing initial threat assessment

 

Module 2: OWASP Top 10 for LLMs - Part 1 (2.5 hours)

Learning Objectives:

  • Master the first five critical LLM vulnerabilities

  • Understand attack vectors and exploitation techniques

  • Apply practical mitigation strategies

Topics Covered:

LLM01: Prompt Injection

  • Direct and indirect prompt injection techniques

  • Hidden instruction attacks and cross-prompt contamination

  • Practical examples: Jailbreaking chatbots and bypassing safety measures

  • Defense strategies: Input sanitization, prompt filtering, differential privacy

LLM02: Sensitive Information Disclosure

  • Training data extraction and system prompt leakage

  • Model behavior analysis for sensitive information exposure

  • Privacy implications and regulatory compliance considerations

  • Mitigation: Output filtering, access controls, data anonymization

LLM03: Supply Chain Vulnerabilities

  • Third-party model dependencies and plugin security

  • Compromised training datasets and model poisoning

  • Vendor risk assessment for AI components

  • Secure model deployment and verification practices

Practical Exercise: Hands-on lab demonstrating prompt injection attacks against vulnerable LLM applications and implementing defensive measures

 

Module 3: OWASP Top 10 for LLMs - Part 2 (2 hours)

Topics Covered:

LLM04: Data and Model Poisoning

  • Training data manipulation techniques

  • Model behavior modification through poisoned inputs

  • Backdoor attacks and data integrity verification

  • Prevention: Data validation pipelines, provenance tracking

LLM05: Improper Output Handling

  • Insecure processing of LLM-generated content

  • Code injection through AI-generated outputs

  • Cross-site scripting via AI responses

  • Output validation and sanitization frameworks

Practical Exercise: Simulating data poisoning attacks and implementing robust output validation mechanisms

 

Module 4: Advanced LLM Threats (1.5 hours)

Topics Covered:

LLM06: Excessive Agency

  • Autonomous decision-making risks and boundary violations

  • Agent authority and permission management

  • Unintended system interactions and privilege escalation

  • Implementing guardrails and human oversight controls

LLM07: System Prompt Leakage

  • System instruction exposure vulnerabilities

  • Credential and logic disclosure through prompts

  • Attack techniques for extracting system prompts

  • Securing system instructions and external configuration

Practical Exercise: Designing secure agent architectures with appropriate access controls and monitoring

 

Day 2: Advanced Threats and Implementation

Module 5: Emerging AI Threats (2 hours)

Learning Objectives:

  • Understand cutting-edge AI security threats

  • Implement advanced detection and prevention techniques

  • Design resilient AI systems against sophisticated attacks

Topics Covered:

LLM08: Vector and Embedding Weaknesses

  • RAG system vulnerabilities and vector database security

  • Embedding poisoning and similarity manipulation attacks

  • Adversarial examples in semantic search

  • Securing vector stores and implementing anomaly detection

LLM09: Misinformation and Model Reliability

  • Hallucination detection and mitigation

  • Bias amplification and fairness considerations

  • Fact-checking and source verification mechanisms

  • Content validation and human oversight integration

LLM10: Unbounded Consumption

  • Resource exhaustion and denial-of-service attacks

  • Rate limiting and resource management strategies

  • Cost optimization and budget controls

  • Performance monitoring and alerting systems

Practical Exercise: Building a secure RAG pipeline with vector database protection and hallucination detection

 

Module 6: Agentic AI Security (2 hours)

Learning Objectives:

  • Understand the unique security challenges of autonomous AI agents

  • Apply the OWASP Agentic AI taxonomy to real-world systems

  • Implement security controls for multi-agent environments

Topics Covered:

  • Introduction to Agentic AI and autonomous systems

  • OWASP Agentic AI Threat Taxonomy: Agent Design, Memory, Planning, Tool Use, Deployment

  • Multi-agent system security and coordination risks

  • Tool misuse, memory poisoning, and goal hijacking attacks

  • Securing agent communication and decision-making processes

Practical Exercise: Threat modeling exercise using OWASP Agentic AI taxonomy on a multi-agent customer service system

 

Module 7: OWASP Threat Defense COMPASS Implementation (2 hours)

Learning Objectives:

  • Master the practical application of Threat Defense COMPASS

  • Integrate AI threat assessment into organizational security programs

  • Develop comprehensive AI risk management strategies

Topics Covered:

  • Deep dive into Threat Defense COMPASS methodology

  • OODA Loop integration: Observe, Orient, Decide, Act

  • Mapping threats to MITRE ATT&CK and ATLAS frameworks

  • Building AI Threat Resilience Strategy Dashboards

  • Integration with existing security tools and processes

Practical Exercise: Complete threat assessment using COMPASS for a Microsoft Copilot deployment scenario

 

Module 8: Practical Implementation and Best Practices (2.5 hours)

Learning Objectives:

  • Design secure AI architectures from the ground up

  • Implement monitoring and incident response for AI systems

  • Create governance frameworks for AI security

Topics Covered:

Secure AI Development Lifecycle:

  • Security-by-design principles for AI applications

  • Code review practices for LLM integrations

  • Testing methodologies and vulnerability scanning

  • Deployment security and production hardening

Monitoring and Detection:

  • AI-specific logging and monitoring requirements

  • Anomaly detection for AI systems

  • Incident response procedures for AI security events

  • Forensics and investigation techniques

Governance and Compliance:

  • AI risk management frameworks and policies

  • Regulatory compliance considerations (GDPR, AI Act, etc.)

  • Third-party risk assessment for AI vendors

  • Security awareness training for AI development teams

Practical Exercise: Design a complete security architecture for an enterprise AI chatbot including monitoring, governance, and incident response procedures

 

Module 9: Tools and Technologies (1 hour)

Learning Objectives:

  • Evaluate and implement AI security tools

  • Understand the current AI security solutions landscape

  • Build practical detection and prevention capabilities

Topics Covered:

  • AI security tool ecosystem and vendor landscape

  • Open-source security tools: Garak, PyRIT, Giskard

  • Commercial solutions for AI security and monitoring

  • Integration patterns and deployment strategies

  • Tool selection criteria and evaluation frameworks

Practical Exercise: Hands-on demonstration of AI security testing tools and implementation planning

 

Module 10: Future Trends and Wrap-up (1 hour)

Learning Objectives:

  • Understand emerging threats and future security challenges

  • Develop continuous learning and improvement strategies

  • Create action plans for organizational AI security programs

Topics Covered:

  • Emerging threats: Deepfakes, advanced prompt injection, model inversion

  • Future OWASP GenAI project developments and roadmap

  • Building AI security communities and knowledge sharing

  • Continuous improvement and threat intelligence integration

Action Planning Exercise: Develop a 90-day action plan for implementing OWASP GenAI security practices in participants' organizations

Requirements

  • General understanding of web application security principles
  • Basic familiarity with AI/ML concepts
  • Experience with security frameworks or risk assessment methodologies preferred

Audience

  • Cybersecurity professionals
  • AI developers
  • System architects
  • Compliance officers
  • Security practitioners
 14 Hours

Delivery Options

Private Group Training

Our identity is rooted in delivering exactly what our clients need.

  • Pre-course call with your trainer
  • Customisation of the learning experience to achieve your goals -
    • Bespoke outlines
    • Practical hands-on exercises containing data / scenarios recognisable to the learners
  • Training scheduled on a date of your choice
  • Delivered online, onsite/classroom or hybrid by experts sharing real world experience

Private Group Prices RRP from €4560 online delivery, based on a group of 2 delegates, €1440 per additional delegate (excludes any certification / exam costs). We recommend a maximum group size of 12 for most learning events.

Contact us for an exact quote and to hear our latest promotions


Public Training

Please see our public courses

Provisional Upcoming Courses (Contact Us For More Information)

Related Categories