AFSD: Adaptive Feature Space Distillation for Distributed Deep Learning
Permanent lenke
https://hdl.handle.net/10037/27512Dato
2022-08-08Type
Journal articleTidsskriftartikkel
Peer reviewed
Forfatter
Khaleghian, Salman; Ullah, Habib; Johnsen, Einar Broch; Andersen, Anders; Marinoni, AndreaSammendrag
We propose a novel and adaptive feature space distillation method (AFSD) to reduce the communication overhead among distributed computers. The proposed method improves the Codistillation process by supporting longer update interval rates. AFSD performs knowledge distillates across the models infrequently and provides flexibility to the models in terms of exploring diverse variations in the training process. We perform knowledge distillation in terms of sharing the feature space instead of output only. Therefore, we also propose a new loss function for the Codistillation technique in AFSD. Using the feature space leads to more efficient knowledge transfer between models with a longer update interval rates. In our method, the models can achieve the same accuracy as Allreduce and Codistillation with fewer epochs.
Er en del av
Khaleghian, S. (2022). Scalable computing for earth observation - Application on Sea Ice analysis. (Doctoral thesis). https://hdl.handle.net/10037/27513.Forlag
IEEESitering
Khaleghian S, Ullah H, Johnsen E. B., Andersen A, Marinoni A. AFSD: Adaptive Feature Space Distillation for Distributed Deep Learning. IEEE Access. 2022;10:84569-84578Metadata
Vis full innførselSamlinger
Copyright 2022 The Author(s)