Annotating 3D point cloud data is labor-intensive. Self-supervised representation learning can reduce the intense demand of manual annotation. However, the sparsity of point cloud, while containing rich geometric structural information, makes the self-supervised representation learning of point clouds more difficult than that of 2D images, especially for point cloud contrastive learning. Recent works employ simple augmentations on point clouds to construct contrastive pairs, but they overlook the geometry structure point cloud data, leading to degraded quality in contrastive views. To compensate the insufficiency in contrasting contrastive pairs, we propose a novel contrastive learning approach to delve deeply into the intrinsic geometric structure of point clouds, termed partial contrastive learning. Specifically, we apply a mask to a portion of the structure in a point cloud sample, while preserving the integrity of the structure in another point cloud. By comparing the structure variations between these point clouds, the model is enabled to encode the geometric information into the self-supervised representations, thereby enhancing the model to maximize the similarity among features that exhibit similar structures. We pretrain our model on the ShapeNet dataset and evaluate its transferability to tasks including classification, segmentation, and few-shot classification. Our method achieves a 90.94% linear SVM accuracy with contrastive training alone, outperforming ToThePoint by 0.91% in point cloud self-supervised learning. Additionally, our method demonstrates superior performance in segmentation and few-shot classification compared to Point-BERT.
Bohua WangZhiqiang TianAixue YeFeng WenShaoyi DuYue Gao
Weichao DingZehao YangFei LuoGu ChunhuaWenbo Dong
Xiaoxiao ShengZhiqiang ShenGang Xiao
Zhuoyang ZhangYuhao DongYunze LiuYi Li
Siming YanZhan-Ying YangHaoxiang LiLi GuanKun HaoGang HuaQixing HuangQixing Huang