Umeå University's logo

umu.sePublications
Change search
CiteExportLink to record
Permanent link

Direct link
Cite
Citation style
  • apa
  • ieee
  • vancouver
  • Other style
More styles
Language
  • de-DE
  • en-GB
  • en-US
  • fi-FI
  • nn-NO
  • nn-NB
  • sv-SE
  • Other locale
More languages
Output format
  • html
  • text
  • asciidoc
  • rtf
Three Operator Splitting with Subgradients, Stochastic Gradients, and Adaptive Learning Rates
Umeå University, Faculty of Science and Technology, Department of Mathematics and Mathematical Statistics. Umeå University.ORCID iD: 0000-0001-7320-1506
Massachusetts Institute of Technology.
Massachusetts Institute of Technology.
2021 (English)In: Advances in Neural Information Processing Systems 34 (NeurIPS 2021) / [ed] M. Ranzato, A. Beygelzimer, P.S. Liang, J.W. Vaughan, Y. Dauphin, San Diego: Neural Information Processing Systems , 2021, p. 19743-19756Conference paper, Published paper (Refereed)
Abstract [en]

Three Operator Splitting (TOS) (Davis & Yin, 2017) can minimize the sum of multiple convex functions effectively when an efficient gradient oracle or proximal operator is available for each term. This requirement often fails in machine learning applications: (i) instead of full gradients only stochastic gradients may be available; and (ii) instead of proximal operators, using subgradients to handle complex penalty functions may be more efficient and realistic. Motivated by these concerns, we analyze three potentially valuable extensions of TOS. The first two permit using subgradients and stochastic gradients, and are shown to ensure a O(1/√t) convergence rate. The third extension AdapTOS endows TOS with adaptive step-sizes. For the important setting of optimizing a convex loss over the intersection of convex sets AdapTOS attains universal convergence rates, i.e., the rate adapts to the unknown smoothness degree of the objective. We compare our proposed methods with competing methods on various applications.

Place, publisher, year, edition, pages
San Diego: Neural Information Processing Systems , 2021. p. 19743-19756
Keywords [en]
Three operator splitting, Davis-Yin splitting, adaptive optimization, nonsmooth optimization, subgradient method, stochastic optimization, proximal methods, proximal gradient method
National Category
Computational Mathematics
Identifiers
URN: urn:nbn:se:umu:diva-190486Scopus ID: 2-s2.0-85131876010OAI: oai:DiVA.org:umu-190486DiVA, id: diva2:1620755
Conference
Thirty-fifth Conference on Neural Information Processing Systems (NeruIPS 2021), Online, 6-14 December, 2021.
Funder
Wallenberg AI, Autonomous Systems and Software Program (WASP)
Note

ISBN: 978-171384539-3

Available from: 2021-12-16 Created: 2021-12-16 Last updated: 2024-07-02Bibliographically approved

Open Access in DiVA

No full text in DiVA

Other links

ScopusPublisher´s fulltext

Authority records

Yurtsever, Alp

Search in DiVA

By author/editor
Yurtsever, Alp
By organisation
Department of Mathematics and Mathematical Statistics
Computational Mathematics

Search outside of DiVA

GoogleGoogle Scholar

urn-nbn

Altmetric score

urn-nbn
Total: 343 hits
CiteExportLink to record
Permanent link

Direct link
Cite
Citation style
  • apa
  • ieee
  • vancouver
  • Other style
More styles
Language
  • de-DE
  • en-GB
  • en-US
  • fi-FI
  • nn-NO
  • nn-NB
  • sv-SE
  • Other locale
More languages
Output format
  • html
  • text
  • asciidoc
  • rtf