JOURNAL ARTICLE

Knowledge distillation based on multi-layer fusion features

Shengyuan TanRongzuo GuoJialiang TangNing JiangJunying Zou

Year: 2023 Journal:   PLoS ONE Vol: 18 (8)Pages: e0285901-e0285901   Publisher: Public Library of Science

Abstract

Knowledge distillation improves the performance of a small student network by promoting it to learn the knowledge from a pre-trained high-performance but bulky teacher network. Generally, most of the current knowledge distillation methods extract relatively simple features from the middle or bottom layer of teacher network for knowledge transmission. However, the above methods ignore the fusion of features, and the fused features contain richer information. We believe that the richer and better information contained in the knowledge delivered by teachers to students, the easier it is for students to perform better. In this paper, we propose a new method called Multi-feature Fusion Knowledge Distillation (MFKD) to extract and utilize the expressive fusion features of teacher network. Specifically, we extract feature maps from different positions of the network, i.e., the middle layer, the bottom layer, and even the front layer of the network. To properly utilize these features, this method designs a multi-feature fusion scheme to integrate them together. Compared to features extracted from single location of teacher network, the final fusion feature map contains meaningful information. Extensive experiments on image classification tasks demonstrate that the student network trained by our MFKD can learn from the fusion features, leading to superior performance. The results show that MFKD can improve the Top-1 accuracy of ResNet20 and VGG8 by 1.82% and 3.35% respectively on the CIFAR-100 dataset, which is better than state-of-the-art many existing methods.

Keywords:
Feature (linguistics) Computer science Distillation Layer (electronics) Fusion Artificial intelligence Artificial neural network Scheme (mathematics) Pattern recognition (psychology) Data mining Feature extraction Machine learning Mathematics

Metrics

4
Cited By
0.73
FWCI (Field Weighted Citation Impact)
58
Refs
0.66
Citation Normalized Percentile
Is in top 1%
Is in top 10%

Citation History

Topics

Advanced Neural Network Applications
Physical Sciences →  Computer Science →  Computer Vision and Pattern Recognition
Brain Tumor Detection and Classification
Life Sciences →  Neuroscience →  Neurology
Domain Adaptation and Few-Shot Learning
Physical Sciences →  Computer Science →  Artificial Intelligence

Related Documents

JOURNAL ARTICLE

Knowledge Distillation in Histology Landscape by Multi-Layer Features Supervision

Sajid JavedArif MahmoodTalha QaiserNaoufel Werghi

Journal:   IEEE Journal of Biomedical and Health Informatics Year: 2023 Vol: 27 (4)Pages: 2037-2046
JOURNAL ARTICLE

Layer-fusion for online mutual knowledge distillation

Gan HuYanli JiXingzhu LiangYuexing Han

Journal:   Multimedia Systems Year: 2022 Vol: 29 (2)Pages: 787-796
JOURNAL ARTICLE

Dual-Layer Fusion Knowledge Reasoning with Enhanced Multi-modal Features

JING Boxiang, WANG Hairong, WANG Tong, YANG Zhenye

Journal:   DOAJ (DOAJ: Directory of Open Access Journals) Year: 2025
JOURNAL ARTICLE

Knowledge Fusion Distillation: Improving Distillation with Multi-scale Attention Mechanisms

Linfeng LiWeixing SuFang LiuMaowei HeXiaodan Liang

Journal:   Neural Processing Letters Year: 2023 Vol: 55 (5)Pages: 6165-6180
© 2026 ScienceGate Book Chapters — All rights reserved.