Federated Learning of Neural Network Models with Heterogeneous Structures

Federated learning trains a model on a centralized server using datasets distributed over a large number of edge devices. Applying federated learning ensures data privacy because it does not transfer local data from edge devices to the server. Existing federated learning algorithms assume that all d...

Full description

Saved in:
Bibliographic Details
Published in:2020 19th IEEE International Conference on Machine Learning and Applications (ICMLA) pp. 735 - 740
Main Authors: Thonglek, Kundjanasith, Takahashi, Keichi, Ichikawa, Kohei, Iida, Hajimu, Nakasan, Chawanat
Format: Conference Proceeding
Language:English
Published: IEEE 01-12-2020
Subjects:
Online Access:Get full text
Tags: Add Tag
No Tags, Be the first to tag this record!
Description
Summary:Federated learning trains a model on a centralized server using datasets distributed over a large number of edge devices. Applying federated learning ensures data privacy because it does not transfer local data from edge devices to the server. Existing federated learning algorithms assume that all deployed models share the same structure. However, it is often infeasible to distribute the same model to every edge device because of hardware limitations such as computing performance and storage space. This paper proposes a novel federated learning algorithm to aggregate information from multiple heterogeneous models. The proposed method uses weighted average ensemble to combine the outputs from each model. The weight for the ensemble is optimized using black box optimization methods. We evaluated the proposed method using diverse models and datasets and found that it can achieve comparable performance to conventional training using centralized datasets. Furthermore, we compared six different optimization methods to tune the weights for the weighted average ensemble and found that tree parzen estimator achieves the highest accuracy among the alternatives.
DOI:10.1109/ICMLA51294.2020.00120