Boosting online feature transfer via separable feature fusion

Lujun Li, Shiuan Ni Liang, Ya Yang, Zhe Jin

Research output: Chapter in Book/Report/Conference proceedingConference PaperResearchpeer-review

7 Citations (Scopus)

Abstract

Feature distillation is a widely used training method to transfer feature information from a teacher to a student network. Current methods seek to minimize the reconstruction error of hidden feature maps between teacher-student models by explicitly optimizing distillation loss. However, some feature loss methods require complex transformations, which are not easy to optimize. In this paper, we propose a novel and effective feature distillation method, which learns to transfer knowledge by applying feature fusion as an alternative to distillation loss. Specifically, we fuse the intermediate feature of the student model to the attention teacher network, which has better representation and relatively less training cost. During training, this separable feature fusion can effectively transfer feature knowledge and is easy to optimize without complex transformation. After training, the feature fusion and the teacher network can be discarded, and the student network can be used separately in inference. Equipped with auxiliary classifier for ensemble logits distillation, our Separable Feature Knowledge Distillation (SFKD) obtains state-of-the-art performance. In experiments, SFKD achieves 4% performance improvement on CIFAR-100 and 2% on ImageNet for ResNet models, which substantially outperforms other feature distillation methods.

Original languageEnglish
Title of host publication2022 International Joint Conference on Neural Networks (IJCNN) - 2022 Conference Proceedings
EditorsMarco Gori, Alessandro Sperduti
Place of PublicationPiscataway NJ USA
PublisherIEEE, Institute of Electrical and Electronics Engineers
Number of pages7
ISBN (Electronic)9781728186719
ISBN (Print)9781665495264
DOIs
Publication statusPublished - 2022
EventIEEE International Joint Conference on Neural Networks 2022 - Padua, Italy
Duration: 18 Jul 202223 Jul 2022
https://ieeexplore.ieee.org/xpl/conhome/9891857/proceeding (Proceedings)

Publication series

NameProceedings of the International Joint Conference on Neural Networks
PublisherIEEE, Institute of Electrical and Electronics Engineers
Volume2022-July
ISSN (Print)2161-4393
ISSN (Electronic)2161-4407

Conference

ConferenceIEEE International Joint Conference on Neural Networks 2022
Abbreviated titleIJCNN 2022
Country/TerritoryItaly
CityPadua
Period18/07/2223/07/22
Internet address

Keywords

  • Features Fused
  • Knowledge Distillation
  • Semantic Matching

Cite this