no code implementations • 8 Apr 2024 • Weikai Lu, Ziqian Zeng, Jianwei Wang, Zhengdong Lu, Zelin Chen, Huiping Zhuang, Cen Chen
Jailbreaking attacks can enable Large Language Models (LLMs) to bypass the safeguard and generate harmful content.
General Knowledge