Convergence and Trade-Offs in Riemannian Gradient Descent and Riemannian Proximal Point

David Martı́nez-Rubio, Christophe Roux, Sebastian Pokutta
Proceedings of the 41st International Conference on Machine Learning, PMLR 235:34920-34948, 2024.

Abstract

In this work, we analyze two of the most fundamental algorithms in geodesically convex optimization: Riemannian gradient descent and (possibly inexact) Riemannian proximal point. We quantify their rates of convergence and produce different variants with several trade-offs. Crucially, we show the iterates naturally stay in a ball around an optimizer, of radius depending on the initial distance and, in some cases, on the curvature. Previous works simply assumed bounded iterates, resulting in rates that were not fully quantified. We also provide an implementable inexact proximal point algorithm and prove several new useful properties of Riemannian proximal methods: they work when positive curvature is present, the proximal operator does not move points away from any optimizer, and we quantify the smoothness of its induced Moreau envelope. Further, we explore beyond our theory with empirical tests.

Cite this Paper


BibTeX
@InProceedings{pmlr-v235-marti-nez-rubio24a, title = {Convergence and Trade-Offs in {R}iemannian Gradient Descent and {R}iemannian Proximal Point}, author = {Mart\'{\i}nez-Rubio, David and Roux, Christophe and Pokutta, Sebastian}, booktitle = {Proceedings of the 41st International Conference on Machine Learning}, pages = {34920--34948}, year = {2024}, editor = {Salakhutdinov, Ruslan and Kolter, Zico and Heller, Katherine and Weller, Adrian and Oliver, Nuria and Scarlett, Jonathan and Berkenkamp, Felix}, volume = {235}, series = {Proceedings of Machine Learning Research}, month = {21--27 Jul}, publisher = {PMLR}, pdf = {https://raw.githubusercontent.com/mlresearch/v235/main/assets/marti-nez-rubio24a/marti-nez-rubio24a.pdf}, url = {https://proceedings.mlr.press/v235/marti-nez-rubio24a.html}, abstract = {In this work, we analyze two of the most fundamental algorithms in geodesically convex optimization: Riemannian gradient descent and (possibly inexact) Riemannian proximal point. We quantify their rates of convergence and produce different variants with several trade-offs. Crucially, we show the iterates naturally stay in a ball around an optimizer, of radius depending on the initial distance and, in some cases, on the curvature. Previous works simply assumed bounded iterates, resulting in rates that were not fully quantified. We also provide an implementable inexact proximal point algorithm and prove several new useful properties of Riemannian proximal methods: they work when positive curvature is present, the proximal operator does not move points away from any optimizer, and we quantify the smoothness of its induced Moreau envelope. Further, we explore beyond our theory with empirical tests.} }
Endnote
%0 Conference Paper %T Convergence and Trade-Offs in Riemannian Gradient Descent and Riemannian Proximal Point %A David Martı́nez-Rubio %A Christophe Roux %A Sebastian Pokutta %B Proceedings of the 41st International Conference on Machine Learning %C Proceedings of Machine Learning Research %D 2024 %E Ruslan Salakhutdinov %E Zico Kolter %E Katherine Heller %E Adrian Weller %E Nuria Oliver %E Jonathan Scarlett %E Felix Berkenkamp %F pmlr-v235-marti-nez-rubio24a %I PMLR %P 34920--34948 %U https://proceedings.mlr.press/v235/marti-nez-rubio24a.html %V 235 %X In this work, we analyze two of the most fundamental algorithms in geodesically convex optimization: Riemannian gradient descent and (possibly inexact) Riemannian proximal point. We quantify their rates of convergence and produce different variants with several trade-offs. Crucially, we show the iterates naturally stay in a ball around an optimizer, of radius depending on the initial distance and, in some cases, on the curvature. Previous works simply assumed bounded iterates, resulting in rates that were not fully quantified. We also provide an implementable inexact proximal point algorithm and prove several new useful properties of Riemannian proximal methods: they work when positive curvature is present, the proximal operator does not move points away from any optimizer, and we quantify the smoothness of its induced Moreau envelope. Further, we explore beyond our theory with empirical tests.
APA
Martı́nez-Rubio, D., Roux, C. & Pokutta, S.. (2024). Convergence and Trade-Offs in Riemannian Gradient Descent and Riemannian Proximal Point. Proceedings of the 41st International Conference on Machine Learning, in Proceedings of Machine Learning Research 235:34920-34948 Available from https://proceedings.mlr.press/v235/marti-nez-rubio24a.html.

Related Material