Adaptive Variant of the Frank-Wolfe Algorithm for Convex Optimization Problems

封面

如何引用文章

全文:

开放存取 开放存取
受限制的访问 ##reader.subscriptionAccessGranted##
受限制的访问 订阅存取

详细

In this paper, a variant of the Frank–Wolfe method for convex optimization problems with adaptive selection of the step parameter corresponding to information about the smoothness of the target function (the Lipschitz constant of the gradient) was investigated. Theoretical estimates of the quality of the approximate solution given out by the method using adaptively selected parameters L_k are obtained. On a class of problems on a convex feasible set with a convex objective function, the guaranteed convergence rate of the proposed method is sublinear. The special subclass of such problems is considered (the objective function with the condition of gradient dominance) and estimate of the convergence rate using adaptively selected parameters L_k is obtained. An important feature of the obtained result is the elaboration of a situation in which it is possible to guarantee, after the completion of the iteration, a reduction of the discrepancy in the function by at least 2 times. At the same time, the use of adaptively selected parameters in theoretical estimates makes it possible to apply the method for both smooth and non-smooth problems, provided that the exit criterion from the iteration is met. For smooth problems, it can be proved that the theoretical estimates of the method are guaranteed to be optimal up to multiplication by a constant factor. Computational experiments were performed, and a comparison with two other algorithms was carried out, during which the efficiency of the algorithm was demonstrated for a number of both smooth and non-smooth problems.

作者简介

G. Aivazian

Moscow Institute of Physics and Technology

编辑信件的主要联系方式.
Email: aivazian.grigory25@yandex.ru
Russia, 141701, Moscow region, Dolgoprudny, Institutskiy per., 9

F. Stonyakin

Moscow Institute of Physics and Technology
; Vernadsky Crimean Federal University

编辑信件的主要联系方式.
Email: fedyor@mail.ru
Russia, 141701, Moscow region, Dolgoprudny, Institutskiy per., 9; Russia, 295007, Simferopol, Academician Vernadsky Avenue, 4

D. Pasechnyk

Moscow Institute of Physics and Technology
; Trusted Artificial Intelligence Research Center of ISP RAS

编辑信件的主要联系方式.
Email: dmivilensky1@gmail.com
Russia, 141701, Moscow region, Dolgoprudny, Institutskiy per., 9; Russia, 109004, Moscow, Alexander Solzhenitsyn st., 25

M. Alkousa

Moscow Institute of Physics and Technology
; National Research University “Higher School of Economics”

编辑信件的主要联系方式.
Email: mohammad.alkousa@phystech.edu
Russia, 141701, Moscow region, Dolgoprudny, Institutskiy per., 9; Russia, 101000, Moscow, Myasnitskaya st., 20

A. Raigorodsky

Moscow Institute of Physics and Technology
; Moscow State University M. V. Lomonosov, Faculty of Mechanics and Mathematics
; Caucasian Mathematical Center of the Adyghe State University

编辑信件的主要联系方式.
Email: raigorodsky@yandex-team.ru
Russia, 141701, Moscow region, Dolgoprudny, Institutskiy per., 9; Russia, 119991, Moscow, Leninskiye Gory, 1; Republic of Adygea, 385016, Maykop, Pervomaiskaya st., 208

I. Baran

Vernadsky Crimean Federal University

编辑信件的主要联系方式.
Email: matemain@mail.ru
Russia, 295007, Simferopol, Academician Vernadsky Avenue, 4

参考

  1. Canon M.D., Cullum C.D. A tight upper bound on the rate of convergence of Frank–Wolfe algorithm // SIAM Journal on Control. 1968. V. 6 (2.4). P. 509–516.
  2. Bomze I.M., Rinaldi F., Zeffiro D. Frank–Wolfe and friends: a journey into projection-free first-order optimization methods // 4OR-Q J Oper Res. 2021. V. 19. P. 313–345.
  3. Braun G., Carderera A., Combettes C.W., Hassani H., Karbasi A., Mokhtari A., Pokutta S. Conditional Gradient Methods. https://arxiv.org/pdf/2211.14103.pdf
  4. Nesterov Y. Complexity bounds for primal-dual methods minimizing the model of objective function // Math. Program. 2018. V. 171 (1-2). P. 311–330.
  5. Nesterov Y. Universal gradient methods for convex optimization problems // Math. Program. A 2015. V. 152. P. 381–404.
  6. Pedregosa F., Negiar G., Askari A., Jaggi M. Linearly convergent Frank–Wolfe with backtracking line-search. In: International Conference on Artificial Intelligence and Statistics. Proceedings of Machine Learning Research. 2020. P. 1–10.
  7. Polyak B.T. Gradient methods for minimizing functionals (in Russian) // Zh. Vychisl. Mat. Mat. Fiz. 1963. P. 643–653.
  8. Łojasiewicz S. A topological property of real analytic subsets (in French) // Coll. du CNRS, Les équations aux dérivos partielles. 1963. P. 87–89.
  9. Karimi H., Nutini J., Schmidt M. Linear convergence of gradient and proximal-gradient methods under the Polyak–Łojasiewicz condition // Machine Learning and Knowledge Discovery in Databases: European Conference, ECML PKDD 2016, Riva del Garda, Italy, September 19–23, 2016, Proceedings, Part I 16. Springer International Publishing, 2016. P. 795–811.
  10. Freund R.M., Grigas P., Mazumder R. An extended Frank–Wolfe method within face directions, and its application to low-rank matrix completion // SIAM Journal on Optimization. 2017. V. 27 (2.1). P. 319–346.
  11. ,000 ratings and 3,600 tag applications applied to 9,000 movies by 600 users. Last updated 9/2018. https://grouplens.org/datasets/movielens/
  12. Vapnik V. The Nature of Statistical Learning Theory. Springer. 2013.
  13. Clarkson K.L. Coresets, sparse greedy approximation, and the Frank–Wolfe algorithm // ACM Transactions on Algorithms. 2010. V. 6 (2.4). P. 1–30.
  14. Pima Indians Diabetes Database. https://www.kaggle.com/datasets/uciml/pima-indians-diabetes-database
  15. Ivanov V.K., Vasin V.V., Tanana V.P. Theory of linear ill-posed problems and its applications. Walter de Gruyter. 2013.
  16. LIBSVM Data: Classification (Binary Class). https://www.csie.ntu.edu.tw/cjlin/libsvmtools/datasets/binary.html
  17. Левитин Е.С., Поляк Б.Т. Методы минимизации при наличии ограничений // Журнал вычислит. матем. и матем. физ. 1966. Т. 6. № 5. С. 787–823.
  18. Candes E.J., Recht B. Exact matrix completion via convex optimization // Foundations of Computational Mathematics. 2009. V. 9 (2.6). P. 717–772.
  19. Combettes C.W., Pokutta S. Complexity of Linear Minimization and Projection on Some Sets // Operations Research Letters. 2021. V. 49 (2.4). P. 565–571.
  20. Frank M., Wolfe P. An algorithm for quadratic programming // Naval Research Logistics Quarterly. 1956. V. 3 (1–2). P. 95–110.

补充文件

附件文件
动作
1. JATS XML
2.

下载 (280KB)
3.

下载 (113KB)
4.

下载 (28KB)
5.

下载 (231KB)
6.

下载 (84KB)
7.

下载 (98KB)
8.

下载 (141KB)

版权所有 © Г.В. Айвазян, Ф.С. Стонякин, Д.А. Пасечнюк, М.С. Алкуса, А.М. Райгородский, И.В. Баран, 2023

##common.cookie##