TY - JOUR
T1 - Efficient Federated Learning with Multi-Teacher Knowledge Distillation for COVID-19 Detection
AU - Annan, Richard
AU - Qin, Hong
AU - Yuan, Xiaohong
AU - Roy, Kaushik
AU - Newman, Robert
AU - Qingge, Letu
PY - 2024/12/16
Y1 - 2024/12/16
N2 - The growing availability of COVID-19 data and advancements in AI offer potential for improved pandemic prediction and prevention. Federated Learning (FL) frameworks support collaborative, privacy-preserving COVID-19 detection, but often neglect the need for simpler models in resource-constrained settings. Knowledge distillation, where a complex “teacher” model transfers insights to a simpler “student” model, struggles to retain detailed information, especially with a single teacher. To address this, we propose a novel FL algorithm, FL-MTKD, which uses multiple teachers to distill knowledge into an efficient 2.5 MB student model. Our results show that while simplified architectures like FL-SimpCNN (2.5 MB) handle non-IID datasets better, larger models like FL-CovidCNN (20.74 MB) and FL-DeepCovid (351.6 MB) perform poorly in such settings. FL-MTKD outperforms other models, achieving 89.74% accuracy and 89.71% F1 score on non-IID datasets, and over 93% accuracy on IID (Independent and Identically Distributed) datasets, offering strong generalization with minimal storage needs. Our developed code can be found from QinggeLab-ACMBCB-24 on github.
AB - The growing availability of COVID-19 data and advancements in AI offer potential for improved pandemic prediction and prevention. Federated Learning (FL) frameworks support collaborative, privacy-preserving COVID-19 detection, but often neglect the need for simpler models in resource-constrained settings. Knowledge distillation, where a complex “teacher” model transfers insights to a simpler “student” model, struggles to retain detailed information, especially with a single teacher. To address this, we propose a novel FL algorithm, FL-MTKD, which uses multiple teachers to distill knowledge into an efficient 2.5 MB student model. Our results show that while simplified architectures like FL-SimpCNN (2.5 MB) handle non-IID datasets better, larger models like FL-CovidCNN (20.74 MB) and FL-DeepCovid (351.6 MB) perform poorly in such settings. FL-MTKD outperforms other models, achieving 89.74% accuracy and 89.71% F1 score on non-IID datasets, and over 93% accuracy on IID (Independent and Identically Distributed) datasets, offering strong generalization with minimal storage needs. Our developed code can be found from QinggeLab-ACMBCB-24 on github.
KW - COVID-19 Detection
KW - Federated Learning
KW - Knowledge Distillation
KW - Medical Imaging
KW - Resource-Constrained Environments
UR - https://www.scopus.com/inward/record.uri?partnerID=HzOxMe3b&scp=85216418495&origin=inward
UR - https://www.scopus.com/inward/citedby.uri?partnerID=HzOxMe3b&scp=85216418495&origin=inward
U2 - 10.1145/3698587.3701396
DO - 10.1145/3698587.3701396
M3 - Conference article
JO - ACM-BCB 2024 - 15th ACM Conference on Bioinformatics, Computational Biology, and Health Informatics
JF - ACM-BCB 2024 - 15th ACM Conference on Bioinformatics, Computational Biology, and Health Informatics
M1 - 56
T2 - 15th ACM Conference on Bioinformatics, Computational Biology, and Health Informatics, ACM-BCB 2024
Y2 - 22 November 2024 through 25 November 2024
ER -