Loading [MathJax]/jax/output/SVG/config.js
Russian Journal of Nonlinear Dynamics
RUS  ENG    JOURNALS   PEOPLE   ORGANISATIONS   CONFERENCES   SEMINARS   VIDEO LIBRARY   PACKAGE AMSBIB  
General information
Latest issue
Archive
Impact factor

Search papers
Search references

RSS
Latest issue
Current issues
Archive issues
What is RSS



Rus. J. Nonlin. Dyn.:
Year:
Volume:
Issue:
Page:
Find






Personal entry:
Login:
Password:
Save password
Enter
Forgotten password?
Register


Russian Journal of Nonlinear Dynamics, 2024, Volume 20, Number 5, Pages 759–788
DOI: https://doi.org/10.20537/nd241209
(Mi nd922)
 

NONLINEAR SYSTEMS IN ROBOTICS

Accelerated Zero-Order SGD under High-Order Smoothness and Overparameterized Regime

G. K. Bychkovab, D. M. Dvinskikhcad, A. V. Antsiferovaea, A. V. Gasnikovfgc, A. V. Lobanovcha

a ISP RAS Research Center for Trusted Artificial Intelligence, ul. Alexandra Solzhenitsyna 25, Moscow, 109004 Russia
b Lomonosov Moscow State University, ul. Leninskiye Gory 1, Moscow, 119991 Russia
c Moscow Institute of Physics and Technology, Institutskiy per. 9, Dolgoprudny, 141701 Russia
d National Research University Higher School of Economics, Pokrovsky bul. 11, Moscow, 109028 Russia
e Institute for Artificial Intelligence, Lomonosov Moscow State University, Lomonosovsky pr. 27/1, Moscow, 119192 Russia
f Innopolis University, ul. Universitetskaya 1, Innopolis, 420500, Russia
g Institute for Information Transmission Problems of RAS, per. Bolshoy Karetny 19, Moscow, 127051 Russia
h Skolkovo Institute of Science and Technology, Bolshoy Boulevard 30, bld. 1, Moscow, 121205 Russia
References:
Abstract: We present a novel gradient-free algorithm to solve a convex stochastic optimization problem, such as those encountered in medicine, physics, and machine learning (e.g., the adversarial multi-armed bandit problem), where the objective function can only be computed through numerical simulation, either as the result of a real experiment or as feedback given by the function evaluations from an adversary. Thus, we suppose that only black-box access to the function values of the objective is available, possibly corrupted by adversarial noise: deterministic or stochastic. The noisy setup can arise naturally from modeling randomness within a simulation or by computer discretization, or when exact values of the function are forbidden due to privacy issues, or when solving nonconvex problems as convex ones with an inexact function oracle. By exploiting higher-order smoothness, fulfilled, e.g., in logistic regression, we improve the performance of zero-order methods developed under the assumption of classical smoothness (or having a Lipschitz gradient). The proposed algorithm enjoys optimal oracle complexity and is designed under an overparameterization setup, i.e., when the number of model parameters is much larger than the size of the training dataset. Overparametrized models fit to the training data perfectly while also having good generalization and outperforming underparameterized models on unseen data. We provide convergence guarantees for the proposed algorithm under both types of noise. Moreover, we estimate the maximum permissible adversarial noise level that maintains the desired accuracy in the Euclidean setup, and then we extend our results to a non-Euclidean setup. Our theoretical results are verified using the logistic regression problem.
Keywords: zero-order optimization, gradient-free algorithms, high-order smoothness, kernel approximation, overparametrization
Funding agency Grant number
Russian Science Foundation 21-71-30005
The research was supported by Russian Science Foundation (project No. 21-71-30005), https://rscf.ru/en/project/21-71-30005/.
Received: 03.11.2024
Accepted: 17.12.2024
Document Type: Article
MSC: 49M30
Language: English
Citation: G. K. Bychkov, D. M. Dvinskikh, A. V. Antsiferova, A. V. Gasnikov, A. V. Lobanov, “Accelerated Zero-Order SGD under High-Order Smoothness and Overparameterized Regime”, Rus. J. Nonlin. Dyn., 20:5 (2024), 759–788
Citation in format AMSBIB
\Bibitem{BycDviAnt24}
\by G. K. Bychkov, D. M. Dvinskikh, A. V. Antsiferova, A. V. Gasnikov, A. V. Lobanov
\paper Accelerated Zero-Order SGD under High-Order Smoothness and Overparameterized Regime
\jour Rus. J. Nonlin. Dyn.
\yr 2024
\vol 20
\issue 5
\pages 759--788
\mathnet{http://mi.mathnet.ru/nd922}
\crossref{https://doi.org/10.20537/nd241209}
Linking options:
  • https://www.mathnet.ru/eng/nd922
  • https://www.mathnet.ru/eng/nd/v20/i5/p759
  • Citing articles in Google Scholar: Russian citations, English citations
    Related articles in Google Scholar: Russian articles, English articles
    Russian Journal of Nonlinear Dynamics
    Statistics & downloads:
    Abstract page:37
    Full-text PDF :10
    References:8
     
      Contact us:
     Terms of Use  Registration to the website  Logotypes © Steklov Mathematical Institute RAS, 2025