LLM-based agents now automate scientific data analysis, but they risk producing plausible yet false claims through selective analysis. A new arXiv paper warns that fluent explanations do not equal verification. Researchers must implement adversarial experiments to validate results. This prevents the rapid generation of publishable positives that lack genuine evidence.