Adversarial Prompting in LLM gggggggggggg

SaahilKarnik1 36 views 7 slides Oct 03, 2024
Slide 1
Slide 1 of 7
Slide 1
1
Slide 2
2
Slide 3
3
Slide 4
4
Slide 5
5
Slide 6
6
Slide 7
7

About This Presentation

Adversarial Prompting in LLMs


Slide Content

Adversarial Prompting in LLMs Saahil Karnik

What is Adversarial Prompting?

Prompt Injection

Prompt Leaking

Jailbreaking (DAN)

Defense against Prompt Injection

References Adversarial Machine Learning: A Taxonomy and Terminology of Attacks and Mitigations (Jan 2024) Can AI really be protected from text-based attacks? (Feb 2023) Prompt injection attacks against GPT-3 (Sep 2022) Adversarial Prompting in LLMs Using GPT Eliezer against ChatGPT Jailbreaking Jailbreak Chat
Tags