Generalized stochastic Frank–Wolfe algorithm with stochastic “substitute” gradient for structured convex optimization
Abstract The stochastic Frank–Wolfe method has recently attracted much general interest in the context of optimization for statistical and machine learning due to its ability to work with a more general feasible region. However, there has been a complexity gap in the dependence on the...
Main Authors: | , |
---|---|
Other Authors: | |
Format: | Article |
Language: | English |
Published: |
Springer Berlin Heidelberg
2021
|
Online Access: | https://hdl.handle.net/1721.1/136776 |
_version_ | 1811077923267936256 |
---|---|
author | Lu, Haihao Freund, Robert M |
author2 | Massachusetts Institute of Technology. Department of Mathematics |
author_facet | Massachusetts Institute of Technology. Department of Mathematics Lu, Haihao Freund, Robert M |
author_sort | Lu, Haihao |
collection | MIT |
description | Abstract
The stochastic Frank–Wolfe method has recently attracted much general interest in the context of optimization for statistical and machine learning due to its ability to work with a more general feasible region. However, there has been a complexity gap in the dependence on the optimality tolerance
$$\varepsilon $$
ε
in the guaranteed convergence rate for stochastic Frank–Wolfe compared to its deterministic counterpart. In this work, we present a new generalized stochastic Frank–Wolfe method which closes this gap for the class of structured optimization problems encountered in statistical and machine learning characterized by empirical loss minimization with a certain type of “linear prediction” property (formally defined in the paper), which is typically present in loss minimization problems in practice. Our method also introduces the notion of a “substitute gradient” that is a not-necessarily-unbiased estimate of the gradient. We show that our new method is equivalent to a particular randomized coordinate mirror descent algorithm applied to the dual problem, which in turn provides a new interpretation of randomized dual coordinate descent in the primal space. Also, in the special case of a strongly convex regularizer our generalized stochastic Frank–Wolfe method (as well as the randomized dual coordinate descent method) exhibits linear convergence. Furthermore, we present computational experiments that indicate that our method outperforms other stochastic Frank–Wolfe methods for a sufficiently small optimality tolerance, consistent with the theory developed herein. |
first_indexed | 2024-09-23T10:50:26Z |
format | Article |
id | mit-1721.1/136776 |
institution | Massachusetts Institute of Technology |
language | English |
last_indexed | 2024-09-23T10:50:26Z |
publishDate | 2021 |
publisher | Springer Berlin Heidelberg |
record_format | dspace |
spelling | mit-1721.1/1367762023-09-27T17:57:59Z Generalized stochastic Frank–Wolfe algorithm with stochastic “substitute” gradient for structured convex optimization Lu, Haihao Freund, Robert M Massachusetts Institute of Technology. Department of Mathematics Sloan School of Management Abstract The stochastic Frank–Wolfe method has recently attracted much general interest in the context of optimization for statistical and machine learning due to its ability to work with a more general feasible region. However, there has been a complexity gap in the dependence on the optimality tolerance $$\varepsilon $$ ε in the guaranteed convergence rate for stochastic Frank–Wolfe compared to its deterministic counterpart. In this work, we present a new generalized stochastic Frank–Wolfe method which closes this gap for the class of structured optimization problems encountered in statistical and machine learning characterized by empirical loss minimization with a certain type of “linear prediction” property (formally defined in the paper), which is typically present in loss minimization problems in practice. Our method also introduces the notion of a “substitute gradient” that is a not-necessarily-unbiased estimate of the gradient. We show that our new method is equivalent to a particular randomized coordinate mirror descent algorithm applied to the dual problem, which in turn provides a new interpretation of randomized dual coordinate descent in the primal space. Also, in the special case of a strongly convex regularizer our generalized stochastic Frank–Wolfe method (as well as the randomized dual coordinate descent method) exhibits linear convergence. Furthermore, we present computational experiments that indicate that our method outperforms other stochastic Frank–Wolfe methods for a sufficiently small optimality tolerance, consistent with the theory developed herein. 2021-11-01T14:33:17Z 2021-11-01T14:33:17Z 2020-03-04 2021-04-21T03:31:56Z Article http://purl.org/eprint/type/JournalArticle https://hdl.handle.net/1721.1/136776 en https://doi.org/10.1007/s10107-020-01480-7 Creative Commons Attribution-Noncommercial-Share Alike http://creativecommons.org/licenses/by-nc-sa/4.0/ Springer-Verlag GmbH Germany, part of Springer Nature and Mathematical Optimization Society application/pdf Springer Berlin Heidelberg Springer Berlin Heidelberg |
spellingShingle | Lu, Haihao Freund, Robert M Generalized stochastic Frank–Wolfe algorithm with stochastic “substitute” gradient for structured convex optimization |
title | Generalized stochastic Frank–Wolfe algorithm with stochastic “substitute” gradient for structured convex optimization |
title_full | Generalized stochastic Frank–Wolfe algorithm with stochastic “substitute” gradient for structured convex optimization |
title_fullStr | Generalized stochastic Frank–Wolfe algorithm with stochastic “substitute” gradient for structured convex optimization |
title_full_unstemmed | Generalized stochastic Frank–Wolfe algorithm with stochastic “substitute” gradient for structured convex optimization |
title_short | Generalized stochastic Frank–Wolfe algorithm with stochastic “substitute” gradient for structured convex optimization |
title_sort | generalized stochastic frank wolfe algorithm with stochastic substitute gradient for structured convex optimization |
url | https://hdl.handle.net/1721.1/136776 |
work_keys_str_mv | AT luhaihao generalizedstochasticfrankwolfealgorithmwithstochasticsubstitutegradientforstructuredconvexoptimization AT freundrobertm generalizedstochasticfrankwolfealgorithmwithstochasticsubstitutegradientforstructuredconvexoptimization |