// Popular Articles

#alignment
#5842025-12-14

Same Model, Same Prompt, Two Answers: How GPU Precision Silently Breaks LLM Safety

A new paper shows aligned LLMs refuse under BF16 and jailbreak under FP16 on the same prompt up to 84% of the time. No safety benchmark is controlling for precision — which means every public refusal rate you trust today is precision-dependent.

llm-safetyalignmentjailbreak
7 phút đọc