A new arXiv paper warns that LLM agents accelerate a failure mode where plausible but false analyses are generated for scientific data. These tools optimize for publishable positives rather than rigorous verification. Researchers argue that adversarial experiments must replace iterative code accumulation. This shift prevents agents from treating hypothesis spaces as mere candidate claims.