JOURNAL ARTICLE

DenseCL: A simple framework for self-supervised dense visual pre-training

Xinlong WangRufeng ZhangChunhua ShenTao Kong

Year: 2022 Journal:   Visual Informatics Vol: 7 (1)Pages: 30-40   Publisher: Elsevier BV

Abstract

Self-supervised learning aims to learn a universal feature representation without labels. To date, most existing self-supervised learning methods are designed and optimized for image classification. These pre-trained models can be sub-optimal for dense prediction tasks due to the discrepancy between image-level prediction and pixel-level prediction. To fill this gap, we aim to design an effective, dense self-supervised learning framework that directly works at the level of pixels (or local features) by taking into account the correspondence between local features. Specifically, we present dense contrastive learning (DenseCL), which implements self-supervised learning by optimizing a pairwise contrastive (dis)similarity loss at the pixel level between two views of input images. Compared to the supervised ImageNet pre-training and other self-supervised learning methods, our self-supervised DenseCL pre-training demonstrates consistently superior performance when transferring to downstream dense prediction tasks including object detection, semantic segmentation and instance segmentation. Specifically, our approach significantly outperforms the strong MoCo-v2 by 2.0% AP on PASCAL VOC object detection, 1.1% AP on COCO object detection, 0.9% AP on COCO instance segmentation, 3.0% mIoU on PASCAL VOC semantic segmentation and 1.8% mIoU on Cityscapes semantic segmentation. The improvements are up to 3.5% AP and 8.8% mIoU over MoCo-v2, and 6.1% AP and 6.1% mIoU over supervised counterpart with frozen-backbone evaluation protocol.Code and models are available at: https://git.io/DenseCL

Keywords:
Pascal (unit) Artificial intelligence Segmentation Pattern recognition (psychology) Computer science Pairwise comparison Supervised learning Feature (linguistics) Object detection Machine learning Artificial neural network

Metrics

8
Cited By
1.57
FWCI (Field Weighted Citation Impact)
54
Refs
0.81
Citation Normalized Percentile
Is in top 1%
Is in top 10%

Citation History

Topics

Domain Adaptation and Few-Shot Learning
Physical Sciences →  Computer Science →  Artificial Intelligence
Multimodal Machine Learning Applications
Physical Sciences →  Computer Science →  Computer Vision and Pattern Recognition
Advanced Image and Video Retrieval Techniques
Physical Sciences →  Computer Science →  Computer Vision and Pattern Recognition

Related Documents

JOURNAL ARTICLE

Object Adaptive Self-Supervised Dense Visual Pre-Training

Yu ZhangTao ZhangHongyuan ZhuZihan ChenSiya MiXi PengXin Geng

Journal:   IEEE Transactions on Image Processing Year: 2025 Vol: 34 Pages: 2228-2240
JOURNAL ARTICLE

UniVIP: A Unified Framework for Self-Supervised Visual Pre-training

Zhaowen LiYousong ZhuFan YangWei LiChaoyang ZhaoYingying ChenZiyang ChenJiahao XieLiwei WuRui ZhaoMing TangJinqiao Wang

Journal:   2022 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR) Year: 2022 Pages: 14607-14616
JOURNAL ARTICLE

3SAT: A Simple Self-Supervised Adversarial Training Framework

Fang JiangHaonan HeJiyan SunJiadong FuZhaorui GuoYinlong LiuWei Ma

Journal:   Proceedings of the AAAI Conference on Artificial Intelligence Year: 2025 Vol: 39 (16)Pages: 16523-16531
© 2026 ScienceGate Book Chapters — All rights reserved.