Generalized self-concordant analysis of Frank–Wolfe algorithms

Loading...
Thumbnail Image
Date
2022
Volume
198
Issue
Journal
Series Titel
Book Title
Publisher
Berlin ; Heidelberg : Springer
Abstract

Projection-free optimization via different variants of the Frank–Wolfe method has become one of the cornerstones of large scale optimization for machine learning and computational statistics. Numerous applications within these fields involve the minimization of functions with self-concordance like properties. Such generalized self-concordant functions do not necessarily feature a Lipschitz continuous gradient, nor are they strongly convex, making them a challenging class of functions for first-order methods. Indeed, in a number of applications, such as inverse covariance estimation or distance-weighted discrimination problems in binary classification, the loss is given by a generalized self-concordant function having potentially unbounded curvature. For such problems projection-free minimization methods have no theoretical convergence guarantee. This paper closes this apparent gap in the literature by developing provably convergent Frank–Wolfe algorithms with standard O(1/k) convergence rate guarantees. Based on these new insights, we show how these sublinearly convergent methods can be accelerated to yield linearly convergent projection-free methods, by either relying on the availability of a local liner minimization oracle, or a suitable modification of the away-step Frank–Wolfe method.

Description
Keywords
Convex programming, Frank–Wolfe algorithm, Generalized self-concordant functions
Citation
Dvurechensky, P., Safin, K., Shtern, S., & Staudigl, M. (2022). Generalized self-concordant analysis of Frank–Wolfe algorithms. 198. https://doi.org//10.1007/s10107-022-01771-1
License
CC BY 4.0 Unported