JOURNAL ARTICLE

One‐stage self‐distillation guided knowledge transfer for long‐tailed visual recognition

Yuelong XiaShu ZhangJun WangWei ZouJuxiang ZhouBin Wen

Year: 2022 Journal:   International Journal of Intelligent Systems Vol: 37 (12)Pages: 11893-11908   Publisher: Wiley

Abstract

Deep learning has achieved remarkable progress for visual recognition on balanced data sets but still performs poorly on real-world long-tailed data distribution. The existing methods mainly decouple the problem into the two-stage decoupling training, that is, representation learning and classifier training, or multistage training based on knowledge distillation, thus resulting in huge training steps and extra computation cost. In this paper, we propose a conceptually simple yet effective One-stage Long-tailed Self-Distillation framework, called OLSD, which simultaneously takes representation learning and classifier training into one-stage training. For representation learning, we take two different sampling distributions and mixup them to input them into two branches, where the collaborative consistency loss is introduced to train network consistency, and we theoretically show that the proposed mixup naturally generates a tail-majority distribution mixup. For classifier training, we introduce balanced self-distillation guided knowledge transfer to improve generalization performance, where we theoretically show that proposed knowledge transfer implicitly minimizes not only cross-entropy but also KL divergence between head-to-tail and tail-to-head. Extensive experiments on long-tailed CIFAR10/100, ImageNet-LT and multilabel long-tailed VOC-LT demonstrate the proposed method's effectiveness.

Keywords:
Computer science Distillation Artificial intelligence Machine learning Classifier (UML) Transfer of learning Pattern recognition (psychology)

Metrics

2
Cited By
0.39
FWCI (Field Weighted Citation Impact)
29
Refs
0.61
Citation Normalized Percentile
Is in top 1%
Is in top 10%

Citation History

Topics

Domain Adaptation and Few-Shot Learning
Physical Sciences →  Computer Science →  Artificial Intelligence
Advanced Image and Video Retrieval Techniques
Physical Sciences →  Computer Science →  Computer Vision and Pattern Recognition
Multimodal Machine Learning Applications
Physical Sciences →  Computer Science →  Computer Vision and Pattern Recognition

Related Documents

BOOK-CHAPTER

Attention-Guided Feature Distillation for Long-Tailed Visual Recognition

Wei HeHai‐Lin LiuLei Chen

Communications in computer and information science Year: 2025 Pages: 422-433
JOURNAL ARTICLE

Self Supervision to Distillation for Long-Tailed Visual Recognition

Tianhao LiLimin WangGangshan Wu

Journal:   2021 IEEE/CVF International Conference on Computer Vision (ICCV) Year: 2021 Pages: 610-619
JOURNAL ARTICLE

Balanced self-distillation for long-tailed recognition

Ning RenXiaosong LiYanxia WuYan Fu

Journal:   Knowledge-Based Systems Year: 2024 Vol: 290 Pages: 111504-111504
© 2026 ScienceGate Book Chapters — All rights reserved.