• Online, Instructor-Led
Course Description

Dive into advanced prompt hacking and learn how to identify, exploit, and mitigate vulnerabilities in Large Language Models (LLMs). From Jailbreak attacks to Prompt Injection and Cognitive Hacking, this course equips you with cutting-edge techniques to stay ahead in the evolving field of AI security.

Learning Objectives

Master advanced techniques for securing AI systems from adversarial attacks.

Framework Connections

The materials within this course focus on the NICE Framework Task, Knowledge, and Skill statements identified within the indicated NICE Framework component(s):