arXiv Analytics

Sign in

arXiv:2003.13563 [cs.LG]AbstractReferencesReviewsResources

Stochastic Flows and Geometric Optimization on the Orthogonal Group

Krzysztof Choromanski, David Cheikhi, Jared Davis, Valerii Likhosherstov, Achille Nazaret, Achraf Bahamou, Xingyou Song, Mrugank Akarte, Jack Parker-Holder, Jacob Bergquist, Yuan Gao, Aldo Pacchiano, Tamas Sarlos, Adrian Weller, Vikas Sindhwani

Published 2020-03-30Version 1

We present a new class of stochastic, geometrically-driven optimization algorithms on the orthogonal group $O(d)$ and naturally reductive homogeneous manifolds obtained from the action of the rotation group $SO(d)$. We theoretically and experimentally demonstrate that our methods can be applied in various fields of machine learning including deep, convolutional and recurrent neural networks, reinforcement learning, normalizing flows and metric learning. We show an intriguing connection between efficient stochastic optimization on the orthogonal group and graph theory (e.g. matching problem, partition functions over graphs, graph-coloring). We leverage the theory of Lie groups and provide theoretical results for the designed class of algorithms. We demonstrate broad applicability of our methods by showing strong performance on the seemingly unrelated tasks of learning world models to obtain stable policies for the most difficult $\mathrm{Humanoid}$ agent from $\mathrm{OpenAI}$ $\mathrm{Gym}$ and improving convolutional neural networks.

Related articles: Most relevant | Search more
arXiv:2108.00051 [cs.LG] (Published 2021-07-30)
Coordinate descent on the orthogonal group for recurrent neural network training
arXiv:1906.02435 [cs.LG] (Published 2019-06-06)
Complete Dictionary Learning via $\ell^4$-Norm Maximization over the Orthogonal Group
arXiv:2106.14662 [cs.LG] (Published 2021-06-25)
Improving Uncertainty Calibration of Deep Neural Networks via Truth Discovery and Geometric Optimization