Understanding Progressive Training Through the Framework of Randomized Coordinate Descent

Rafal Szlendak, Elnur Gasanov, Peter Richtárik

Research output: Contribution to conferencePaperpeer-review

Abstract

We propose a Randomized Progressive Training algorithm (RPT) – a stochastic proxy for the well-known Progressive Training method (PT) (Karras et al., 2018). Originally designed to train GANs (Goodfellow et al., 2014), PT was proposed as a heuristic, with no convergence analysis even for the simplest objective functions. On the contrary, to the best of our knowledge, RPT is the first PT-type algorithm with rigorous and sound theoretical guarantees for general smooth objective functions. We cast our method into the established framework of Randomized Coordinate Descent (RCD) (Nesterov, 2012; Richtárik and Takáč, 2014), for which (as a by-product of our investigations) we also propose a novel, simple and general convergence analysis encapsulating strongly-convex, convex and nonconvex objectives. We then use this framework to establish a convergence theory for RPT. Finally, we validate the effectiveness of our method through extensive computational experiments.

Original languageEnglish (US)
Pages2161-2169
Number of pages9
StatePublished - 2024
Event27th International Conference on Artificial Intelligence and Statistics, AISTATS 2024 - Valencia, Spain
Duration: May 2 2024May 4 2024

Conference

Conference27th International Conference on Artificial Intelligence and Statistics, AISTATS 2024
Country/TerritorySpain
CityValencia
Period05/2/2405/4/24

ASJC Scopus subject areas

  • Artificial Intelligence
  • Software
  • Control and Systems Engineering
  • Statistics and Probability

Fingerprint

Dive into the research topics of 'Understanding Progressive Training Through the Framework of Randomized Coordinate Descent'. Together they form a unique fingerprint.

Cite this