Umeå University's logo

umu.sePublications
Change search
CiteExportLink to record
Permanent link

Direct link
Cite
Citation style
  • apa
  • ieee
  • modern-language-association-8th-edition
  • vancouver
  • Other style
More styles
Language
  • de-DE
  • en-GB
  • en-US
  • fi-FI
  • nn-NO
  • nn-NB
  • sv-SE
  • Other locale
More languages
Output format
  • html
  • text
  • asciidoc
  • rtf
Advancing federated learning: algorithms and use-cases
Umeå University, Faculty of Science and Technology, Department of Computing Science. (Autonomous Distributed Systems Lab)ORCID iD: 0000-0002-3451-2851
2024 (English)Doctoral thesis, comprehensive summary (Other academic)Alternative title
Förbättrad federerad maskininlärning : algoritmer och tillämpningar (Swedish)
Abstract [en]

Federated Learning (FL) is a distributed machine learning paradigm that enables the training of models across numerous clients or organizations without requiring the transfer of local data. This method addresses concerns about data privacy and ownership by keeping raw data on the client itself and only sharing model updates with a central server. Despite its benefits, federated learning faces unique challenges, such as data heterogeneity, computation and communication overheads, and the need for personalized models. Thereby results in reduced model performance, lower efficiency, and longer training times.

This thesis investigates these issues from theoretical, empirical, and practical application perspectives with four-fold contributions, such as federated feature selection, adaptive client selection, model personalization, and socio-cognitive applications. Firstly, we addressed the data heterogeneity problems for federated feature selection in horizontal FL by developing algorithms based on mutual information and multi-objective optimization. Secondly, we tackled system heterogeneity issues that involved variations in computation, storage, and communication capabilities among clients. We proposed a solution that ranks clients with multi-objective optimization for efficient, fair, and adaptive participation in model training. Thirdly, we addressed the issue of client drift caused by data heterogeneity in hierarchical federated learning with a personalized federated learning approach. Lastly, we focused on two key applications that benefit from the FL framework but suffer from data heterogeneity issues. The first application attempts to predict the level of autobiographic memory recall of events associated with the lifelog image by developing clustered personalized FL algorithms, which help in selecting effective lifelog image cues for cognitive interventions for the clients. The second application is the development of a personal image privacy advisor for each client. Along with data heterogeneity, the privacy advisor faces data scarcity issues. We developed a daisy chain-enabled clustered personalized FL algorithm, which predicts whether an image should be shared, kept private, or recommended for sharing by a third party.

Our findings reveal that the proposed methods significantly outperformed the current state-of-the-art FL  algorithms. Our methods deliver superior performance, earlier convergence, and training efficiency.

Place, publisher, year, edition, pages
Umeå: Umeå University, 2024. , p. 84
Series
Report / UMINF, ISSN 0348-0542 ; 24.09
Keywords [en]
Federated Learning, Federated Feature Selection, Statistical Heterogeneity, System Heterogeneity, Model Personalization, Socio-Cognitive Applications
National Category
Computer Sciences
Research subject
Computer Science
Identifiers
URN: urn:nbn:se:umu:diva-228863ISBN: 978-91-8070-463-2 (print)ISBN: 978-91-8070-464-9 (electronic)OAI: oai:DiVA.org:umu-228863DiVA, id: diva2:1892766
Public defence
2024-09-23, Hörsal HUM.D.210, Humanisthuset, Umeå, 13:00 (English)
Opponent
Supervisors
Funder
Wallenberg AI, Autonomous Systems and Software Program (WASP)Available from: 2024-09-02 Created: 2024-08-27 Last updated: 2024-08-28Bibliographically approved
List of papers
1. Fed-FiS: A Novel Information-Theoretic Federated Feature Selection for Learning Stability
Open this publication in new window or tab >>Fed-FiS: A Novel Information-Theoretic Federated Feature Selection for Learning Stability
2021 (English)In: Neural Information Processing: 28th International Conference, ICONIP 2021, Sanur, Bali, Indonesia, December 8–12, 2021, Proceedings, Part V / [ed] Teddy Mantoro, Minho Lee, Media Anugerah Ayu, Kok Wai Wong, Achmad Nizar Hidayanto, Springer Nature, 2021, Vol. 1516, p. 480-487Conference paper, Published paper (Refereed)
Abstract [en]

In the era of big data and federated learning, traditional feature selection methods show unacceptable performance for handling heterogeneity when deployed in federated environments. We propose Fed-FiS, an information-theoretic federated feature selection approach to overcome the problem occur due to heterogeneity. Fed-FiS estimates feature-feature mutual information (FFMI) and feature-class mutual information (FCMI) to generate a local feature subset in each user device. Based on federated values across features and classes obtained from each device, the central server ranks each feature and generates a global dominant feature subset. We show that our approach can find stable features subset collaboratively from all local devices. Extensive experiments based on multiple benchmark iid (independent and identically distributed) and non-iid datasets demonstrate that Fed-FiS significantly improves overall performance in comparison to the state-of-the-art methods. This is the first work on feature selection in a federated learning system to the best of our knowledge.

Place, publisher, year, edition, pages
Springer Nature, 2021
Series
Communications in Computer and Information Science (CCIS), ISSN 1865-0929, E-ISSN 1865-0937 ; 1516
Keywords
Federated learning, Feature selection, Mutual information, Classification, Statistical heterogeneity
National Category
Computer Sciences Computer Systems
Research subject
Computer Science
Identifiers
urn:nbn:se:umu:diva-190137 (URN)10.1007/978-3-030-92307-5_56 (DOI)2-s2.0-85121934752 (Scopus ID)978-3-030-92306-8 (ISBN)978-3-030-92307-5 (ISBN)
Conference
ICONIP: International Conference on Neural Information Processing, Virtual, December 8-12, 2021
Funder
Wallenberg AI, Autonomous Systems and Software Program (WASP), 570011342
Available from: 2021-12-06 Created: 2021-12-06 Last updated: 2024-08-27Bibliographically approved
2. Optimized and adaptive federated learning for straggler-resilient device selection
Open this publication in new window or tab >>Optimized and adaptive federated learning for straggler-resilient device selection
2022 (English)In: 2022 International Joint Conference on Neural Networks (IJCNN), IEEE, 2022, p. 1-9Conference paper, Published paper (Refereed)
Abstract [en]

Federated Learning (FL) has evolved as a promising distributed learning paradigm in which data samples are disseminated over massively connected devices in an IID (Identical and Independent Distribution) or non-IID manner. FL follows a collaborative training approach where each device uses local training data to train local models, and the server generates a global model by combining the local model's parameters. However, FL is vulnerable to system heterogeneity when local devices have varying computational, storage, and communication capabilities over time. The presence of stragglers or low-performing devices in the learning process severely impacts the scalability of FL algorithms and significantly delays convergence. To mitigate this problem, we propose Fed-MOODS, a Multi-Objective Optimization-based Device Selection approach to reduce the effect of stragglers in the FL process. The primary criteria for optimization are to maximize: (i) the availability of the processing capacity of each device, (ii) the availability of the memory in devices, and (iii) the bandwidth capacity of the participating devices. The multi-objective optimization prioritizes devices from fast to slow. The approach involves faster devices in early global rounds and gradually incorporating slower devices from the Pareto fronts to improve the model's accuracy. The overall training time of Fed-MOODS is 1.8× and 1.48× faster than the baseline model (FedAvg) with random device selection for MNIST and FMNIST non-IID data, respectively. Fed-MOODS is extensively evaluated under multiple experimental settings, and the results show that Fed-MOODS has significantly improved model's convergence and performance. Fed-MOODS maintains fairness in the prioritized participation of devices and the model for both IID and non-IID settings.

Place, publisher, year, edition, pages
IEEE, 2022
Series
Proceedings of International Joint Conference on Neural Networks, ISSN 2161-4393, E-ISSN 2161-4407
National Category
Robotics
Identifiers
urn:nbn:se:umu:diva-199974 (URN)10.1109/IJCNN55064.2022.9892777 (DOI)000867070907014 ()2-s2.0-85140800465 (Scopus ID)
Conference
2022 International Joint Conference on Neural Networks (IJCNN), Padua, Italy, July 18-23, 2022
Funder
Wallenberg AI, Autonomous Systems and Software Program (WASP)
Available from: 2022-10-03 Created: 2022-10-03 Last updated: 2024-08-27Bibliographically approved
3. Cost-efficient feature selection for horizontal federated learning
Open this publication in new window or tab >>Cost-efficient feature selection for horizontal federated learning
2024 (English)In: IEEE Transactions on Artificial Intelligence, E-ISSN 2691-4581Article in journal (Refereed) Epub ahead of print
Abstract [en]

Horizontal Federated Learning exhibits substantial similarities in feature space across distinct clients. However, not all features contribute significantly to the training of the global model. Moreover, the curse of dimensionality delays the training. Therefore, reducing irrelevant and redundant features from the feature space makes training faster and inexpensive. This work aims to identify the common feature subset from the clients in federated settings. We introduce a hybrid approach called Fed-MOFS 1 , utilizing Mutual Information and Clustering for local feature selection at each client. Unlike the Fed-FiS, which uses a scoring function for global feature ranking, Fed-MOFS employs multi-objective optimization to prioritize features based on their higher relevance and lower redundancy. This paper compares the performance of Fed-MOFS 2 with conventional and federated feature selection methods. Moreover, we tested the scalability, stability, and efficacy of both Fed-FiS and Fed-MOFS across diverse datasets. We also assessed how feature selection influenced model convergence and explored its impact in scenarios with data heterogeneity. Our results show that Fed-MOFS enhances global model performance with a 50% reduction in feature space and is at least twice as fast as the FSHFL method. The computational complexity for both approaches is O( d 2 ), which is lower than the state-of-the-art.

Place, publisher, year, edition, pages
Institute of Electrical and Electronics Engineers (IEEE), 2024
Keywords
Feature extraction, Computational modeling, Data models, Training, Federated learning, Artificial intelligence, Servers, Clustering, Horizontal Federated Learning, Feature Selection, Mutual Information, Multi-objective Optimization
National Category
Computer Sciences
Identifiers
urn:nbn:se:umu:diva-228215 (URN)10.1109/TAI.2024.3436664 (DOI)2-s2.0-85200235298 (Scopus ID)
Funder
Wallenberg AI, Autonomous Systems and Software Program (WASP)
Available from: 2024-08-05 Created: 2024-08-05 Last updated: 2024-08-27
4. Personalized multi-tier federated learning
Open this publication in new window or tab >>Personalized multi-tier federated learning
2024 (English)Conference paper, Published paper (Refereed)
Abstract [en]

The key challenge of personalized federated learning (PerFL) is to capture the statistical heterogeneity properties of data with inexpensive communications and gain customized performance for participating devices. To address these, we introduced personalized federated learning in multi-tier architecture (PerMFL) to obtain optimized and personalized local models when there are known team structures across devices. We provide theoretical guarantees of PerMFL, which offers linear convergence rates for smooth strongly convex problems and sub-linear convergence rates for smooth non-convex problems. We conduct numerical experiments demonstrating the robust empirical performance of PerMFL, outperforming the state-of-the-art in multiple personalized federated learning tasks.

National Category
Computer Sciences
Identifiers
urn:nbn:se:umu:diva-228859 (URN)
Conference
ICONIP 2024, 31st International Conference on Neural Information Processing, Auckland, New Zealand, December 2-6, 2024
Funder
Wallenberg AI, Autonomous Systems and Software Program (WASP)
Available from: 2024-08-27 Created: 2024-08-27 Last updated: 2024-08-27
5. Predicting event memorability using personalized federated learning
Open this publication in new window or tab >>Predicting event memorability using personalized federated learning
(English)Manuscript (preprint) (Other academic)
Abstract [en]

Lifelog images are very useful as memory cues for recalling past events. Estimating the level of event memory recall induced by a given lifelog image (event memorability) is useful for selecting images for cognitive interventions. Previous works for predicting event memorability follow a centralized model training paradigm that requires several users to share their lifelog images. This risks violating the privacy of individual lifeloggers. Alternatively, a personal model trained with a lifelogger’s own data guarantees privacy. However, it imposes significant effort on the lifelogger to provide a large enough sample of self-rated images to develop a well-performing model for event memorability. Therefore, we propose a clustered personalized federated learning setup, FedMEM, that avoids sharing raw images but still enables collaborative learning via model sharing. For an enhanced learning performance in the presence of data heterogeneity, FedMEM evaluates similarity among users to group them into clusters. We demonstrate that our approach furnishes high-performing personalized models compared to the state-of-the-art.

National Category
Computer Sciences
Identifiers
urn:nbn:se:umu:diva-228858 (URN)
Funder
Wallenberg AI, Autonomous Systems and Software Program (WASP)
Note

The manuscript is under review.

Available from: 2024-08-27 Created: 2024-08-27 Last updated: 2024-08-27
6. The case for federated learning in developing personalized image privacy advisor
Open this publication in new window or tab >>The case for federated learning in developing personalized image privacy advisor
Show others...
(English)Manuscript (preprint) (Other academic)
Abstract [en]

Sharing privacy-sensitive information poses a significant risk when private information is carelessly exposed to individuals or groups through various media. With minimal annotation and user preference, users can decide whether to share or keep an image based on its privacy score. This approach helps reduce the likelihood of privacy breaches and makes it easier to share images with friends or on social media platforms. Federated Learning (FL) allows collaborative learning of models without sharing personal data, which is crucial for sharing privacy-sensitive information. However, there are several challenges associated with learning from private data. This paper proposes two FL algorithms, namely Dynamic-Clustered-FedDC and Apriori-Clustered-FedDC, to address key problems, such as data scarcity, data heterogeneity, and model complexity, in order to ensure personalized image privacy. Both algorithms train personalized models for each annotator using clustered federated learning to address data heterogeneity. Additionally, both algorithms utilize daisy-chaining-based knowledge sharing between annotators to mitigate data scarcity issues during training. In addition to these two algorithms, we propose the PIONet model, which is 20× lighter compared to baseline models and retains equivalent performance. PIONet, along with Dynamic-Clustered-FedDC and Apriori-Clustered-FedDC, outperformed the state-of-the-art federated learning algorithms and the baseline.

National Category
Computer Sciences
Research subject
computational linguistics
Identifiers
urn:nbn:se:umu:diva-228862 (URN)
Funder
Wallenberg AI, Autonomous Systems and Software Program (WASP)
Note

The manuscript is under review

Available from: 2024-08-27 Created: 2024-08-27 Last updated: 2024-08-27

Open Access in DiVA

fulltext(2139 kB)72 downloads
File information
File name FULLTEXT02.pdfFile size 2139 kBChecksum SHA-512
4b542b12b52446eba48ec57b3669563d1758523c41a05f856d6c36288046a9ffc29b420e1d2c9bce811245f4132b42b9d55b4b084bc3638a9894720b6e269c12
Type fulltextMimetype application/pdf
spikblad(182 kB)26 downloads
File information
File name SPIKBLAD01.pdfFile size 182 kBChecksum SHA-512
e1a5b5615dcd371ecebd457515589df269a97354929856da8f92efaca502aa45a140434840b63494235f590b016584388087f30e7f1b5300b47cf16f51f827a7
Type spikbladMimetype application/pdf

Authority records

Banerjee, Sourasekhar

Search in DiVA

By author/editor
Banerjee, Sourasekhar
By organisation
Department of Computing Science
Computer Sciences

Search outside of DiVA

GoogleGoogle Scholar
Total: 72 downloads
The number of downloads is the sum of all downloads of full texts. It may include eg previous versions that are now no longer available

isbn
urn-nbn

Altmetric score

isbn
urn-nbn
Total: 317 hits
CiteExportLink to record
Permanent link

Direct link
Cite
Citation style
  • apa
  • ieee
  • modern-language-association-8th-edition
  • vancouver
  • Other style
More styles
Language
  • de-DE
  • en-GB
  • en-US
  • fi-FI
  • nn-NO
  • nn-NB
  • sv-SE
  • Other locale
More languages
Output format
  • html
  • text
  • asciidoc
  • rtf