The Aurora optimizer introduces leverage-aware updates to handle rectangular matrices more efficiently than standard methods. It specifically addresses gradient variance in non-square weight tensors. This technical refinement improves convergence stability for specific layer architectures. Researchers can now optimize wide or tall layers without the typical performance degradation seen in traditional optimizers.