Rohan's Bytes
Subscribe
Sign in
Jigsaw Puzzles: Splitting Harmful Questions…
Rohan Paul
Nov 7, 2024
Smart LLM-jailbreak attackers can bypass LLM safety by breaking bad requests into harmless pieces.
Read →
Comments
This site requires JavaScript to run correctly. Please
turn on JavaScript
or unblock scripts
Jigsaw Puzzles: Splitting Harmful Questions…
Smart LLM-jailbreak attackers can bypass LLM safety by breaking bad requests into harmless pieces.