Umeå University's logo

umu.sePublications
12345674 of 12
CiteExportLink to record
Permanent link

Direct link
Cite
Citation style
  • apa
  • ieee
  • vancouver
  • Other style
More styles
Language
  • de-DE
  • en-GB
  • en-US
  • fi-FI
  • nn-NO
  • nn-NB
  • sv-SE
  • Other locale
More languages
Output format
  • html
  • text
  • asciidoc
  • rtf
From accelerated first-order methods to structured nonconvex optimization: analysis and perspectives
Umeå University, Faculty of Science and Technology, Department of Mathematics and Mathematical Statistics.ORCID iD: 0000-0001-8251-2605
2026 (English)Doctoral thesis, comprehensive summary (Other academic)Alternative title
Från accelererade första ordningens metoder till strukturerad icke-konvex optimering : analys och perspektiv (Swedish)
Abstract [en]

Modern technologies like machine learning and data-driven decision systems, depend on solving large and often highly complex optimization problems. These problems rarely come with simple shapes or smooth surfaces; instead, they can twist and bend in ways that make finding good solutions surprisingly difficult. This thesis explores two ideas that help us navigate such complexity more effectively. 

The first idea focuses on acceleration and investigates how optimization algorithms can reach good solutions faster while using only basic information such as function values and gradients. By studying these algorithms through a continuous-time analysis, we show that many of the fastest methods behave like carefully designed dynamical systems. This perspective not only clarifies why acceleration happens, but also allows us to design fast algorithms and understand how they behave in the presence of noisy information. 

The second idea focuses on a mathematical structural assumption called difference-of- convex (DC) programming, which captures a remarkably wide range of nonconvex problems. By leveraging this structure, we develop practical algorithms that avoid costly operations like projections or high dimensional gradient evaluations, making them efficient for large-scale applications. Through the connection between DC programming and the classical Expectation–Maximization (EM) algorithm, we develop more scalable EM variants and provide the first general performance guarantees for them. 

Place, publisher, year, edition, pages
Umeå: Umeå University, 2026. , p. 43
Series
Research report in mathematical statistics, ISSN 1653-0829 ; 80/26
Keywords [en]
accelerated methods, convex optimization, nonconvex optimization, DC pro- gramming, EM algorithm, randomized DC algorithm
National Category
Computational Mathematics Probability Theory and Statistics
Identifiers
URN: urn:nbn:se:umu:diva-248866ISBN: 978-91-8070-910-1 (print)ISBN: 978-91-8070-911-8 (electronic)OAI: oai:DiVA.org:umu-248866DiVA, id: diva2:2031159
Public defence
2026-02-24, BIO.E.203 - Aula Biologica + Zoom, 08:30 (English)
Opponent
Supervisors
Funder
Wallenberg AI, Autonomous Systems and Software Program (WASP)
Note

Link to participate via Zoom: https://umu.zoom.us/j/63834231447

Available from: 2026-02-03 Created: 2026-01-22 Last updated: 2026-01-22Bibliographically approved
List of papers
1. A variational perspective on high-resolution ODEs
Open this publication in new window or tab >>A variational perspective on high-resolution ODEs
2023 (English)In: Advances in Neural Information Processing Systems 36 (NeurIPS 2023), Neural information processing systems foundation , 2023Conference paper, Published paper (Refereed)
Abstract [en]

We consider unconstrained minimization of smooth convex functions. We propose a novel variational perspective using forced Euler-Lagrange equation that allows for studying high-resolution ODEs. Through this, we obtain a faster convergence rate for gradient norm minimization using Nesterov's accelerated gradient method. Additionally, we show that Nesterov's method can be interpreted as a rate-matching discretization of an appropriately chosen high-resolution ODE. Finally, using the results from the new variational perspective, we propose a stochastic method for noisy gradients. Several numerical experiments compare and illustrate our stochastic algorithm with state of the art methods.

Place, publisher, year, edition, pages
Neural information processing systems foundation, 2023
Series
Advances in neural information processing systems, ISSN 1049-5258
National Category
Computational Mathematics
Identifiers
urn:nbn:se:umu:diva-223951 (URN)2-s2.0-85191188553 (Scopus ID)
Conference
37th Conference on Neural Information Processing Systems, NeurIPS 2023, New Orleans, USA, December 10-16, 2023
Available from: 2024-05-03 Created: 2024-05-03 Last updated: 2026-01-22Bibliographically approved
2. A unified model for high-resolution ODEs: new insights on accelerated methods
Open this publication in new window or tab >>A unified model for high-resolution ODEs: new insights on accelerated methods
2025 (English)Manuscript (preprint) (Other academic)
Abstract [en]

Recent work on high-resolution ordinary differential equations (HR-ODEs) captures fine nuances among different momentum-based optimization methods, leading to accurate theoretical insights. However, these HR-ODEs often appear disconnected, each targeting a specific algorithm and derived with different assumptions and techniques. We present a unifying framework by showing that these diverse HR-ODEs emerge as special cases of a general HR-ODE derived using the Forced Euler-Lagrange equation. Discretizing this model recovers a wide range of optimization algorithms through different parameter choices. Using integral quadratic constraints, we also introduce a general Lyapunov function to analyze the convergence of the proposed HR-ODE and its discretizations, achieving significant improvements across various cases, including new guarantees for the triple momentum method s HR-ODE and the quasi-hyperbolic momentum method, as well as faster gradient norm minimization rates for Nesterov s accelerated gradient algorithm, among other advances.

National Category
Computational Mathematics
Research subject
Mathematics
Identifiers
urn:nbn:se:umu:diva-248743 (URN)10.48550/arXiv.2503.15136 (DOI)
Available from: 2026-01-22 Created: 2026-01-22 Last updated: 2026-01-22Bibliographically approved
3. Revisiting Frank-Wolfe for structured nonconvex optimization
Open this publication in new window or tab >>Revisiting Frank-Wolfe for structured nonconvex optimization
2025 (English)In: NeuriPS 2025: Downloads, 2025Conference paper, Poster (with or without abstract) (Refereed)
Abstract [en]

We introduce a new projection-free (Frank-Wolfe) method for optimizing structured nonconvex functions that are expressed as a difference of two convex functions. This problem class subsumes smooth nonconvex minimization, positioning our method as a promising alternative to the classical Frank-Wolfe algorithm. DC decompositions are not unique; by carefully selecting a decomposition, we can better exploit the problem structure, improve computational efficiency, and adapt to the underlying problem geometry to find better local solutions. We prove that the proposed method achieves a first-order stationary point in  iterations, matching the complexity of the standard Frank-Wolfe algorithm for smooth non-convex minimization in general. Specific decompositions can, for instance, yield a gradient-efficient variant that requires only  calls to the gradient oracle by reusing computed gradients over multiple iterations. Finally, we present numerical experiments demonstrating the effectiveness of the proposed method compared to other projection-free algorithms.

National Category
Computational Mathematics
Research subject
Mathematics
Identifiers
urn:nbn:se:umu:diva-248742 (URN)
Conference
NeuriPS 2025, The Thirty-Ninth Annual Conference on Neural Information Processing Systems, San Diego, USA, December 1, 2025
Available from: 2026-01-20 Created: 2026-01-20 Last updated: 2026-01-22Bibliographically approved
4. Randomized block coordinate DC algorithm
Open this publication in new window or tab >>Randomized block coordinate DC algorithm
2025 (English)In: EURO Journal on Computational Optimization, ISSN 2192-4406, Vol. 13, article id 100123Article in journal (Refereed) Published
Abstract [en]

We introduce an extension of the Difference of Convex Algorithm (DCA) in the form of a randomized block coordinate approach for problems with separable structure. For n coordinate-blocks and k iterations, our main result proves a non-asymptotic convergence rate of O(n/k) in expectation, with respect to a stationarity measure based on a Forward-Backward envelope. Furthermore, leveraging the connection between DCA and Expectation Maximization (EM), we propose a randomized block coordinate EM algorithm.

Place, publisher, year, edition, pages
Elsevier, 2025
Keywords
Block coordinate methods, Difference of convex, Nonconvex optimization
National Category
Computational Mathematics
Identifiers
urn:nbn:se:umu:diva-247450 (URN)10.1016/j.ejco.2025.100123 (DOI)2-s2.0-105023553755 (Scopus ID)
Funder
Wallenberg AI, Autonomous Systems and Software Program (WASP)
Available from: 2025-12-12 Created: 2025-12-12 Last updated: 2026-01-22Bibliographically approved

Open Access in DiVA

spikblad(234 kB)17 downloads
File information
File name SPIKBLAD01.pdfFile size 234 kBChecksum SHA-512
6bd07218b5607e726cc889ac1ead09999e66241dd7d496eea2dbc447c4be6df1ae389839df5de2379967ee7c31df6115e825cc38d9ce7afede778240f5f4a124
Type spikbladMimetype application/pdf
fulltext(2381 kB)41 downloads
File information
File name FULLTEXT01.pdfFile size 2381 kBChecksum SHA-512
31fe6b0a1c399919fc69ecfb8f8dcd2623956fa5b054ab3d6fba748b024221e6c22d5dccaeffce283e7d4b242b9255e68b4dda54133e13f3aef5cac6dbbcfbaa
Type fulltextMimetype application/pdf

Authority records

Maskan, Hoomaan

Search in DiVA

By author/editor
Maskan, Hoomaan
By organisation
Department of Mathematics and Mathematical Statistics
Computational MathematicsProbability Theory and Statistics

Search outside of DiVA

GoogleGoogle Scholar
The number of downloads is the sum of all downloads of full texts. It may include eg previous versions that are now no longer available

isbn
urn-nbn

Altmetric score

isbn
urn-nbn
Total: 927 hits
12345674 of 12
CiteExportLink to record
Permanent link

Direct link
Cite
Citation style
  • apa
  • ieee
  • vancouver
  • Other style
More styles
Language
  • de-DE
  • en-GB
  • en-US
  • fi-FI
  • nn-NO
  • nn-NB
  • sv-SE
  • Other locale
More languages
Output format
  • html
  • text
  • asciidoc
  • rtf