Umeå University's logo

umu.sePublications
Change search
CiteExportLink to record
Permanent link

Direct link
Cite
Citation style
  • apa
  • ieee
  • vancouver
  • Other style
More styles
Language
  • de-DE
  • en-GB
  • en-US
  • fi-FI
  • nn-NO
  • nn-NB
  • sv-SE
  • Other locale
More languages
Output format
  • html
  • text
  • asciidoc
  • rtf
Equivariance versus augmentation for spherical images
Department of Mathematical Sciences, Chalmers University of Technology, Gothenburg, Sweden; Machine Learning Group at Berlin Institute of Technology, Berlin, Germany; Berlin Institute for the Foundations of Learning and Data (BIFOLD), Berlin, Germany.
Department of Mathematical Sciences, Chalmers University of Technology, Gothenburg, Sweden.
Department of Physics, University of Gothenburg, Gothenburg, Sweden.
Umeå University, Faculty of Science and Technology, Department of Mathematics and Mathematical Statistics.ORCID iD: 0000-0002-3165-6999
Show others and affiliations
2022 (English)In: Proceedings of Machine Learning Research: International Conference on Machine Learning, 17-23 July 2022, Baltimore, Maryland, USA, 2022, Vol. 162, p. 7404-7421Conference paper, Published paper (Refereed)
Abstract [en]

We analyze the role of rotational equivariance in convolutional neural networks (CNNs) applied to spherical images. We compare the performance of the group equivariant networks known as S2CNNs and standard non-equivariant CNNs trained with an increasing amount of data augmentation. The chosen architectures can be considered baseline references for the respective design paradigms. Our models are trained and evaluated on single or multiple items from the MNIST- or FashionMNIST dataset projected onto the sphere. For the task of image classification, which is inherently rotationally invariant, we find that by considerably increasing the amount of data augmentation and the size of the networks, it is possible for the standard CNNs to reach at least the same performance as the equivariant network. In contrast, for the inherently equivariant task of semantic segmentation, the non-equivariant networks are consistently outperformed by the equivariant networks with significantly fewer parameters. We also analyze and compare the inference latency and training times of the different networks, enabling detailed tradeoff considerations between equivariant architectures and data augmentation for practical problems.

Place, publisher, year, edition, pages
2022. Vol. 162, p. 7404-7421
Series
Proceedings of Machine Learning Research, ISSN 2640-3498
National Category
Geometry Computational Mathematics
Research subject
Mathematics
Identifiers
URN: urn:nbn:se:umu:diva-204121Scopus ID: 2-s2.0-85145961098OAI: oai:DiVA.org:umu-204121DiVA, id: diva2:1731592
Conference
39th International Conference on Machine Learning (ICML2022), Baltimore, Maryland, USA, July 17-23, 2022
Funder
Wallenberg AI, Autonomous Systems and Software Program (WASP)Knut and Alice Wallenberg FoundationSwedish Research CouncilAvailable from: 2023-01-27 Created: 2023-01-27 Last updated: 2025-10-24Bibliographically approved

Open Access in DiVA

No full text in DiVA

Other links

ScopusPublisher's full text

Authority records

Ohlsson, Fredrik

Search in DiVA

By author/editor
Ohlsson, Fredrik
By organisation
Department of Mathematics and Mathematical Statistics
GeometryComputational Mathematics

Search outside of DiVA

GoogleGoogle Scholar

urn-nbn

Altmetric score

urn-nbn
Total: 381 hits
CiteExportLink to record
Permanent link

Direct link
Cite
Citation style
  • apa
  • ieee
  • vancouver
  • Other style
More styles
Language
  • de-DE
  • en-GB
  • en-US
  • fi-FI
  • nn-NO
  • nn-NB
  • sv-SE
  • Other locale
More languages
Output format
  • html
  • text
  • asciidoc
  • rtf