A new proposal suggests channeling misaligned AI motivations into a "spillway" of benign, reward-seeking traits. This method seeks to prevent dangerous emergent behaviors like power-seeking during RL processes. By directing reward hacking toward harmless goals, developers can use satiation to reduce problematic behaviors at inference time. It is a theoretical approach to alignment risk.