This paper was converted on www.awesomepapers.org from LaTeX by an anonymous user.
Want to know more? Visit the Converter page.

FedMEKT: Distillation-based Embedding Knowledge Transfer for Multimodal Federated Learning

Huy Q. Le [email protected] Minh N. H. Nguyen [email protected] Chu Myaet Thwal [email protected] Yu Qiao [email protected] Chaoning Zhang [email protected] Choong Seon Hong [email protected]
Abstract

Federated learning (FL) enables a decentralized machine learning paradigm for multiple clients to collaboratively train a generalized global model without sharing their private data. Most existing works have focused on designing FL systems for unimodal data, limiting their potential to exploit valuable multimodal data for future personalized applications. Moreover, the majority of FL approaches still rely on labeled data at the client side, which is often constrained by the inability of users to self-annotate their data in real-world applications. In light of these limitations, we propose a novel multimodal FL framework based on a semi-supervised learning approach to leverage representations from different modalities. To address the challenges of modality discrepancy and labeled data constraints in existing FL systems, our proposed FedMEKT comprises local multimodal autoencoder learning, generalized multimodal autoencoder construction, and generalized classifier learning. Bringing this concept into a system, we develop a distillation-based multimodal embedding knowledge transfer mechanism, namely FedMEKT, which allows the server and clients to exchange joint knowledge extracted from a multimodal proxy dataset. Specifically, our FedMEKT iteratively updates the generalized global encoders with joint embedding knowledge from participating clients through upstream and downstream multimodal embedding knowledge transfer. Through extensive experiments on three multimodal human activity recognition datasets,we demonstrate that FedMEKT not only achieves superior global encoder performance in linear evaluation but also guarantees user privacy for personal data and model parameters while demanding less communication cost than other baselines.

keywords:
\affiliation

[label1]organization=Department of Computer Science and Engineering, Kyung Hee University,city=Yongin-si, postcode=17104, country=Republic of Korea \affiliation[label2]organization=Department of Artificial Intelligence, Kyung Hee University,city=Yongin-si, postcode=17104, country=Republic of Korea \affiliation[label3]organization=Digital Science and Technology Institute, The University of Danang—Vietnam-Korea University of Information and Communication Technology,city=Da Nang, postcode=550000, country=Vietnam

1