Google AI Research released a study testing 12 LLMs for behavioral alignment. The paper measures traits such as honesty, politeness, and risk aversion across prompts. Findings show wide variance, with some models consistently misaligning on safety‑critical tasks. Developers should benchmark these dispositions before real‑world deployments to avoid unintended behavior in production.