The website rlvrbook.com hosts a new mini-book detailing Reinforcement Learning from Verifiable Rewards. This resource focuses on training models using objective, verifiable feedback rather than subjective human preferences. It offers a technical primer for researchers. Practitioners can use these methods to reduce reward hacking in complex alignment tasks.