Google DeepMind introduced Decoupled DiLoCo to enable distributed training across unstable networks. This method separates local gradient updates from global synchronization, preventing a single slow node from stalling the entire cluster. It allows researchers to train large models across geographically distant data centers. Practitioners can now utilize fragmented compute resources without sacrificing training stability.