Publication Date
1-26-2026
Abstract
This paper presents an exploratory case study detailed as a pedagogical experience report on integrating adversarial Large Language Model (LLM) scenarios into a graduate cybersecurity curriculum. In addition to prompt injection, sophisticated techniques such as jailbreaking and model inversion pose emerging threats that traditional computer security curricula often lack. We present the design and implementation of a structured, hands-on module addressing this gap, utilizing a custom Retrieval-Augmented Generation (RAG) platform with local open-source LLMs. A cohort of 16 graduate students participated in this two-week pilot module, engaging in "red team" activities to actively exploit model alignment and privacy vulnerabilities. The module achieved an average post-module quiz score of 88%, and 90% of students reported increased confidence, demonstrating measurable learning outcomes. This report illustrates instructional strategies for translating complex LLM exploits into accessible educational exercises, providing an example educators may adapt to prepare future professionals for the challenges of securing real-world AI systems.
Included in
Educational Methods Commons, Information Security Commons, Scholarship of Teaching and Learning Commons, Technology and Innovation Commons