A new paper on arXiv examines why human annotators disagree on AI safety policies. Researchers distinguish between operational failures, policy ambiguity, and value pluralism to improve data quality. The study suggests that treating all disagreements as simple errors ignores critical nuances. This framework helps developers refine safety guidelines by identifying exactly where human interpretation fails.