Stochastic intermediate gradient method for convex optimization problems


Cite item

Full Text

Open Access Open Access
Restricted Access Access granted
Restricted Access Subscription Access

Abstract

New first-order methods are introduced for solving convex optimization problems from a fairly broad class. For composite optimization problems with an inexact stochastic oracle, a stochastic intermediate gradient method is proposed that allows using an arbitrary norm in the space of variables and a prox-function. The mean rate of convergence of this method and the probability of large deviations from this rate are estimated. For problems with a strongly convex objective function, a modification of this method is proposed and its rate of convergence is estimated. The resulting estimates coincide, up to a multiplicative constant, with lower complexity bounds for the class of composite optimization problems with an inexact stochastic oracle and for all usually considered subclasses of this class.

About the authors

A. V. Gasnikov

Institute for Information Transmission Problems

Author for correspondence.
Email: gasnikov@yandex.ru
Russian Federation, Bol’shoi Karetnyi per. 19/1, Moscow, 127994

P. E. Dvurechensky

Institute for Information Transmission Problems

Email: gasnikov@yandex.ru
Russian Federation, Bol’shoi Karetnyi per. 19/1, Moscow, 127994

Supplementary files

Supplementary Files
Action
1. JATS XML

Copyright (c) 2016 Pleiades Publishing, Ltd.