IBM used a curated dataset of 10 trillion tokens to train Granite 4.1. The team focused on high-quality synthetic data and rigorous filtering to improve reasoning and coding tasks. This architecture emphasizes efficiency for enterprise deployments. Practitioners can now leverage these open-weights models to reduce hallucination rates in specialized business workflows.