Unknown

Dataset Information

0

Beyond convexity-Contraction and global convergence of gradient descent.


ABSTRACT: This paper considers the analysis of continuous time gradient-based optimization algorithms through the lens of nonlinear contraction theory. It demonstrates that in the case of a time-invariant objective, most elementary results on gradient descent based on convexity can be replaced by much more general results based on contraction. In particular, gradient descent converges to a unique equilibrium if its dynamics are contracting in any metric, with convexity of the cost corresponding to the special case of contraction in the identity metric. More broadly, contraction analysis provides new insights for the case of geodesically-convex optimization, wherein non-convex problems in Euclidean space can be transformed to convex ones posed over a Riemannian manifold. In this case, natural gradient descent converges to a unique equilibrium if it is contracting in any metric, with geodesic convexity of the cost corresponding to contraction in the natural metric. New results using semi-contraction provide additional insights into the topology of the set of optimizers in the case when multiple optima exist. Furthermore, they show how semi-contraction may be combined with specific additional information to reach broad conclusions about a dynamical system. The contraction perspective also easily extends to time-varying optimization settings and allows one to recursively build large optimization structures out of simpler elements. Extensions to natural primal-dual optimization and game-theoretic contexts further illustrate the potential reach of these new perspectives.

SUBMITTER: Wensing PM 

PROVIDER: S-EPMC7402485 | biostudies-literature | 2020

REPOSITORIES: biostudies-literature

altmetric image

Publications

Beyond convexity-Contraction and global convergence of gradient descent.

Wensing Patrick M PM   Slotine Jean-Jacques JJ  

PloS one 20200804 8


This paper considers the analysis of continuous time gradient-based optimization algorithms through the lens of nonlinear contraction theory. It demonstrates that in the case of a time-invariant objective, most elementary results on gradient descent based on convexity can be replaced by much more general results based on contraction. In particular, gradient descent converges to a unique equilibrium if its dynamics are contracting in any metric, with convexity of the cost corresponding to the spe  ...[more]

Similar Datasets

| S-EPMC7714173 | biostudies-literature
| S-EPMC8025800 | biostudies-literature
| S-EPMC8563972 | biostudies-literature
| S-EPMC6611597 | biostudies-literature
| S-EPMC2933224 | biostudies-literature
| S-EPMC6923147 | biostudies-literature
| S-EPMC7039878 | biostudies-literature
| S-EPMC10542513 | biostudies-literature
| S-EPMC9198964 | biostudies-literature
| S-EPMC10019646 | biostudies-literature