Abstract: Large Language Models (LLMs) are vulnerable to deceptive jailbreak attacks inducing harmful outputs. Existing defenses suffer from catastrophic forgetting in continual defense learning ...