Large-Language-Models
Injected Approval: A Low Effort Local LLM Jailbreak
A quick look into into one of the simplest attacks on LLM safety mitigations, revealing large gaps in current approaches from major tech companies.
Leaderboard of Open LLMs Ranked by LLM Judges
An evaluation of recent consumer-grade open LLMs based on ratings generated through an LLM-as-a-judge framework.
Evaluating LLM Performance via LLM Judges
Methodology details for how LLMs can rate the performance of other LLMs.