Human goals are inherently under-determined and easily manipulated. This creates a fundamental tension for AI alignment researchers trying to define helpful or obedient systems. Distinguishing between beneficial counsel and harmful brainwashing remains an unsolved technical challenge. Practitioners must now reconcile these unstable human desires with the rigid logic of reward functions.