Skip to main content

Jailbreaking

Injected Approval: A Low Effort Local LLM Jailbreak
Large-Language-Models Jailbreaking Cybersecurity
A quick look into into one of the simplest attacks on LLM safety mitigations, revealing large gaps in current approaches from major tech companies.