Volume 46, Issue 6 pp. 7108-7124
RESEARCH ARTICLE

Incremental subgradient algorithms with dynamic step sizes for separable convex optimizations

Dan Yang

Dan Yang

College of Mathematics and Statistics, Guizhou University, Guiyang, 550025 China

Search for more papers by this author
Xiangmei Wang

Corresponding Author

Xiangmei Wang

College of Mathematics and Statistics, Guizhou University, Guiyang, 550025 China

Correspondence

Xiangmei Wang, College of Mathematics and Statistics, Guizhou University, Guiyang 550025, China.

Email: [email protected]

Communicated by: B. Harrach

Search for more papers by this author
First published: 18 December 2022

Funding information: Research of the second author is supported in part by the National Natural Science Foundation of China (grant 12161017) and Guizhou Provincial Natural Science Foundation of China (grant ZK[2022]110).

Abstract

We consider the incremental subgradient algorithm employing dynamic step sizes for minimizing the sum of a large number of component convex functions. The dynamic step size rule was firstly introduced by Goffin and Kiwiel [Math. Program., 1999, 85(1): 207-211] for the subgradient algorithm, soon later, for the incremental subgradient algorithm by Nedic and Bertsekas in [SIAM J. Optim., 2001, 12(1): 109-138]. It was observed experimentally that the incremental approach has been very successful in solving large separable optimizations and that the dynamic step sizes generally have better computational performance than others in the literature. In the present paper, we propose two modified dynamic step size rules for the incremental subgradient algorithm and analyse the convergence and complexity properties of them. At last, the assignment problem is considered and the incremental subgradient algorithms employing different kinds of dynamic step sizes are applied to solve the problem. The computational experiments show that the two modified ones converges dramatically faster and more stable than the corresponding one in [SIAM J. Optim., 2001, 12(1): 109-138]. Particularly, for solving large separable convex optimizations, we strongly recommend the second one (see Algorithm 3.3 in the paper) since it has interesting computational performance and is the simplest one.

CONFLICT OF INTEREST

This work does not have any conflicts of interest.

The full text of this article hosted at iucr.org is unavailable due to technical difficulties.