The Aurora optimizer introduces leverage-aware scaling to handle rectangular matrices more effectively. This approach reduces gradient noise and stabilizes convergence during training. Researchers claim it outperforms standard Adam in specific high-dimensional settings. Practitioners can expect faster training cycles for models with extreme aspect ratios in their weight matrices.