Breaking LLM safety filters by splitting dangerous prompts into harmless-looking sequential questions.
Share this post
Multi-round jailbreak attack on large…
Share this post
Breaking LLM safety filters by splitting dangerous prompts into harmless-looking sequential questions.