Abstract
The recent leaps in complexity and fluency of Large Language Models (LLMs) mean that, for the first time in human history, people can interact with computers using natural language alone. This creates monumental possibilities of automation and accessibility of computing, but also raises severe security and safety threats: When everyone can interact with LLMs, everyone can potentially break into the systems running LLMs. All it takes is creative use of language. This paper presents Hacc-Man, a game which challenges its players to "jailbreak"an LLM: subvert the LLM to output something that it is not intended to. Jailbreaking is at the intersection between creative problem solving and LLM security. The purpose of the game is threefold: 1. To heighten awareness of the risks of deploying fragile LLMs in everyday systems, 2. To heighten people's self-efficacy in interacting with LLMs, and 3. To discover the creative problem solving strategies, people deploy in this novel context.
Original language | English |
---|---|
Title of host publication | Companion Publication of the 2024 ACM Designing Interactive Systems Conference |
Number of pages | 4 |
Publisher | Association for Computing Machinery |
Publication date | 2024 |
Pages | 338-341 |
DOIs | |
Publication status | Published - 2024 |
Keywords
- LLM security
- arcade games
- red teaming
- jailbreaking
- hacking
- creativity, creative problem solving