sorry-bench/ft-mistral-7b-instruct-v0.2-sorry-bench-202406 Text Generation • Updated Jul 2 • 2.63k • 4
SORRY-Bench: Systematically Evaluating Large Language Model Safety Refusal Behaviors Paper • 2406.14598 • Published Jun 20
Assessing the Brittleness of Safety Alignment via Pruning and Low-Rank Modifications Paper • 2402.05162 • Published Feb 7 • 1
BaDExpert: Extracting Backdoor Functionality for Accurate Backdoor Input Detection Paper • 2308.12439 • Published Aug 23, 2023
Fine-tuning Aligned Language Models Compromises Safety, Even When Users Do Not Intend To! Paper • 2310.03693 • Published Oct 5, 2023 • 1