#5842025-12-14
Same Model, Same Prompt, Two Answers: How GPU Precision Silently Breaks LLM Safety
A new paper shows aligned LLMs refuse under BF16 and jailbreak under FP16 on the same prompt up to 84% of the time. No safety benchmark is controlling for precision — which means every public refusal rate you trust today is precision-dependent.