Jailbreaking
Injected Approval: A Low Effort Local LLM Jailbreak
Large-Language-Models
Jailbreaking
Cybersecurity
A quick look into into one of the simplest attacks on LLM safety mitigations, revealing large gaps in current approaches from major tech companies.