|
MATHEMATICAL MODELING AND NUMERICAL SIMULATION
Variance reduction for minimax problems with a small dimension of one of the variables
E. L. Gladinabc, E. D. Borodichb a Humboldt University of Berlin,
6 Unter den Linden, Berlin, 10117, Germany
b Moscow Institute of Physics and Technology,
9 Institutskiy per., Dolgoprudny, 141701, Russia
c Institute for Information Transmission Problems RAS,
19/1 Bolshoy Karetny per., Moscow, 127051, Russia
Abstract:
The paper is devoted to convex-concave saddle point problems where the objective is a sum of a large number of functions. Such problems attract considerable attention of the mathematical community due to the variety of applications in machine learning, including adversarial learning, adversarial attacks and robustreinforcement learning, to name a few. The individual functions in the sum usually represent losses related to examples from a data set. Additionally, the formulation admits a possibly nonsmooth composite term. Such terms often reflect regularization in machine learning problems. We assume that the dimension of one of the variable groups is relatively small (about a hundred or less), and the other one is large. This case arises, for example, when one considers the dual formulation for a minimization problem with a moderate number of constraints. The proposed approach is based on using Vaidya's cutting plane method to minimize with respect to the outer block of variables. This optimization algorithm is especially effective when the dimension of the problem is not very large. An inexact oracle for Vaidya's method is calculated via an approximate solution of the inner maximization problem, which is solved by the accelerated variance reduced algorithm Katyusha. Thus, we leverage the structure of the problem to achieve fast convergence. Separate complexity bounds for gradients of different components with respect to different variables are obtained in the study. The proposed approach is imposing very mild assumptions about the objective. In particular, neither strong convexity nor smoothness is required with respect to the low-dimensional variable group. The number of steps of the proposed algorithmas well as the arithmetic complexity of each step explicitly depend on the dimensionality of the outer variable, hence the assumption that it is relatively small.
Keywords:
saddle point problems, first-order methods, cutting-plane methods, variance reduction.
Received: 13.02.2022
Citation:
E. L. Gladin, E. D. Borodich, “Variance reduction for minimax problems with a small dimension of one of the variables”, Computer Research and Modeling, 14:2 (2022), 257–275
Linking options:
https://www.mathnet.ru/eng/crm967 https://www.mathnet.ru/eng/crm/v14/i2/p257
|
Statistics & downloads: |
Abstract page: | 126 | Full-text PDF : | 45 | References: | 32 |
|