Rubrics as Rewards (RaR): Training AI to Better Align with Human Preferences Introduction: The Challenge of Training AI for Subjective Tasks When training AI systems to handle complex tasks like medical diagnosis or scientific analysis, we face a fundamental challenge: how do we teach models to produce high-quality outputs when there’s no single “correct” answer? Traditional reinforcement learning methods rely on either: Verifiable rewards (e.g., math problems with clear solutions) Human preference rankings (e.g., scoring multiple responses) But real-world domains like healthcare and science often require balancing objective facts with subjective quality (clarity, completeness, safety). This creates three key problems: …