Abstract
Feature distillation is a widely used training method to transfer feature information from a teacher to a student network. Current methods seek to minimize the reconstruction error of hidden feature maps between teacher-student models by explicitly optimizing distillation loss. However, some feature loss methods require complex transformations, which are not easy to optimize. In this paper, we propose a novel and effective feature distillation method, which learns to transfer knowledge by applying feature fusion as an alternative to distillation loss. Specifically, we fuse the intermediate feature of the student model to the attention teacher network, which has better representation and relatively less training cost. During training, this separable feature fusion can effectively transfer feature knowledge and is easy to optimize without complex transformation. After training, the feature fusion and the teacher network can be discarded, and the student network can be used separately in inference. Equipped with auxiliary classifier for ensemble logits distillation, our Separable Feature Knowledge Distillation (SFKD) obtains state-of-the-art performance. In experiments, SFKD achieves 4% performance improvement on CIFAR-100 and 2% on ImageNet for ResNet models, which substantially outperforms other feature distillation methods.
Original language | English |
---|---|
Title of host publication | 2022 International Joint Conference on Neural Networks (IJCNN) - 2022 Conference Proceedings |
Editors | Marco Gori, Alessandro Sperduti |
Place of Publication | Piscataway NJ USA |
Publisher | IEEE, Institute of Electrical and Electronics Engineers |
Number of pages | 7 |
ISBN (Electronic) | 9781728186719 |
ISBN (Print) | 9781665495264 |
DOIs | |
Publication status | Published - 2022 |
Event | IEEE International Joint Conference on Neural Networks 2022 - Padua, Italy Duration: 18 Jul 2022 → 23 Jul 2022 https://ieeexplore.ieee.org/xpl/conhome/9891857/proceeding (Proceedings) |
Publication series
Name | Proceedings of the International Joint Conference on Neural Networks |
---|---|
Publisher | IEEE, Institute of Electrical and Electronics Engineers |
Volume | 2022-July |
ISSN (Print) | 2161-4393 |
ISSN (Electronic) | 2161-4407 |
Conference
Conference | IEEE International Joint Conference on Neural Networks 2022 |
---|---|
Abbreviated title | IJCNN 2022 |
Country/Territory | Italy |
City | Padua |
Period | 18/07/22 → 23/07/22 |
Internet address |
Keywords
- Features Fused
- Knowledge Distillation
- Semantic Matching