LLM Security
Game Theory Jailbreaking - A New Black-Box Attack Paradigm for LLMs
Jailbreaking Isn’t a Prompting Problem For a long time, jailbreaks were treated as a curiosity. Someone found a clever prompt, it circulated online, a filter was adjusted, and the incident was written off as an edge case. If you’ve spent years in application security, this likely felt familiar