Rohan's Bytes
Subscribe
Sign in
Share this post
Rohan's Bytes
Jigsaw Puzzles: Splitting Harmful Questions to Jailbreak Large Language Models
Copy link
Facebook
Email
Notes
More
Jigsaw Puzzles: Splitting Harmful Questions…
Rohan Paul
Nov 7, 2024
Share this post
Rohan's Bytes
Jigsaw Puzzles: Splitting Harmful Questions to Jailbreak Large Language Models
Copy link
Facebook
Email
Notes
More
Smart LLM-jailbreak attackers can bypass LLM safety by breaking bad requests into harmless pieces.
Read →
Comments
Share
Copy link
Facebook
Email
Notes
More
This site requires JavaScript to run correctly. Please
turn on JavaScript
or unblock scripts
Share this post
Jigsaw Puzzles: Splitting Harmful Questions…
Share this post
Smart LLM-jailbreak attackers can bypass LLM safety by breaking bad requests into harmless pieces.