Umeå University's logo

umu.sePublications
Change search
CiteExportLink to record
Permanent link

Direct link
Cite
Citation style
  • apa
  • ieee
  • vancouver
  • Other style
More styles
Language
  • de-DE
  • en-GB
  • en-US
  • fi-FI
  • nn-NO
  • nn-NB
  • sv-SE
  • Other locale
More languages
Output format
  • html
  • text
  • asciidoc
  • rtf
Deep perceptual similarity is adaptable to ambiguous contexts
Umeå University, Faculty of Science and Technology, Department of Computing Science. (Explainable AI)ORCID iD: 0000-0003-0100-4030
Luleå University of Technology. (Machine Learning)
Luleå University of Technology. (Machine Learning)ORCID iD: 0000-0003-4029-6574
2024 (English)In: Proceedings of the 5th Northern Lights Deep Learning Conference (NLDL), PMLR 233, 2024 / [ed] Tetiana Lutchyn; Adín Ramírez Rivera; Benjamin Ricaud, Cambridge: PMLR , 2024, Vol. 233, p. 212-219Conference paper, Published paper (Refereed)
Abstract [en]

This work examines the adaptability of Deep Perceptual Similarity (DPS) metrics to context beyond those that align with average human perception and contexts in which the standard metrics have been shown to perform well. Prior works have shown that DPS metrics are good at estimating human perception of similarity, so-called perceptual similarity. However, it remains unknown whether such metrics can be adapted to other contexts. In this work, DPS metrics are evaluated for their adaptability to different contradictory similarity contexts. Such contexts are created by randomly ranking six image distortions. Metrics are adapted to consider distortions more or less disruptive to similarity depending on their place in the random rankings. This is done by training pretrained CNNs to measure similarity according to given contexts. The adapted metrics are also evaluated on a perceptual similarity dataset to evaluate whether adapting to a ranking affects their prior performance. The findings show that DPS metrics can be adapted with high performance. While the adapted metrics have difficulties with the same contexts as baselines, performance is improved in 99% of cases. Finally, it is shown that the adaption is not significantly detrimental to prior performance on perceptual similarity. The implementation of this work is available online.

Place, publisher, year, edition, pages
Cambridge: PMLR , 2024. Vol. 233, p. 212-219
Series
Proceedings of Machine Learning Research, E-ISSN 2640-3498 ; 233
Keywords [en]
deep features, perceptual similarity, similarity metrics, computer vision, image similarity, image quality assessment, ambiguity
National Category
Computer graphics and computer vision
Identifiers
URN: urn:nbn:se:umu:diva-221642Scopus ID: 2-s2.0-85189301791OAI: oai:DiVA.org:umu-221642DiVA, id: diva2:1841687
Conference
5th Northern Lights Deep Learning Conference 2024, Tromsø, Norway, 9-11 January 2024.
Available from: 2024-02-29 Created: 2024-02-29 Last updated: 2025-02-07Bibliographically approved

Open Access in DiVA

fulltext(1469 kB)59 downloads
File information
File name FULLTEXT01.pdfFile size 1469 kBChecksum SHA-512
3659fc9ab055d9bf0d581200622a4ca636b31f9c1774c762396aaeb6d00dfc8146a1148bac55c82a5a8fc9632df00e6814181be752911081b02b3fb3040bced9
Type fulltextMimetype application/pdf

Other links

ScopusConference proceeding

Authority records

Pihlgren, Gustav Grund

Search in DiVA

By author/editor
Pihlgren, Gustav GrundLiwicki, Marcus
By organisation
Department of Computing Science
Computer graphics and computer vision

Search outside of DiVA

GoogleGoogle Scholar
Total: 59 downloads
The number of downloads is the sum of all downloads of full texts. It may include eg previous versions that are now no longer available

urn-nbn

Altmetric score

urn-nbn
Total: 242 hits
CiteExportLink to record
Permanent link

Direct link
Cite
Citation style
  • apa
  • ieee
  • vancouver
  • Other style
More styles
Language
  • de-DE
  • en-GB
  • en-US
  • fi-FI
  • nn-NO
  • nn-NB
  • sv-SE
  • Other locale
More languages
Output format
  • html
  • text
  • asciidoc
  • rtf