Five popular AI chatbots provided inaccurate or incomplete medical information in roughly 50% of tested cases. Researchers led by Nick Tiller identified these failures across diverse health queries. The findings highlight a persistent reliability gap in clinical applications. Practitioners should maintain strict human oversight to prevent patient harm from automated misinformation.