GUARD: Role-playing for Testing LLMs with Jailbreaks
The author proposes GUARD, a role-playing system to generate jailbreaks for testing Large Language Models (LLMs) adherence to guidelines. By utilizing four different roles, GUARD aims to improve the safety and reliability of LLM-based applications.