JOURNAL ARTICLE

WhitenedCSE: Whitening-based Contrastive Learning of Sentence Embeddings

Abstract

This paper presents a whitening-based contrastive learning method for sentence embedding learning (WhitenedCSE), which combines contrastive learning with a novel shuffled group whitening. Generally, contrastive learning pulls distortions of a single sample (i.e., positive samples) close and push negative samples far away, correspondingly facilitating the alignment and uniformity in the feature space. A popular alternative to the “pushing” operation is whitening the feature space, which scatters all the samples for uniformity. Since the whitening and the contrastive learning have large redundancy w.r.t. the uniformity, they are usually used separately and do not easily work together. For the first time, this paper integrates whitening into the contrastive learning scheme and facilitates two benefits. 1) Better uniformity. We find that these two approaches are not totally redundant but actually have some complementarity due to different uniformity mechanism. 2) Better alignment. We randomly divide the feature into multiple groups along the channel axis and perform whitening independently within each group. By shuffling the group division, we derive multiple distortions of a single sample and thus increase the positive sample diversity. Consequently, using multiple positive samples with enhanced diversity further improves contrastive learning due to better alignment. Extensive experiments on seven semantic textual similarity tasks show our method achieves consistent improvement over the contrastive learning baseline and sets new states of the art, e.g., 78.78% (+2.53% based on BERT{pasted macro ‘BA’}) Spearman correlation on STS tasks.

Keywords:
Computer science Redundancy (engineering) Artificial intelligence Sentence Embedding Feature (linguistics) Sample (material) Correlation Natural language processing Feature vector Pattern recognition (psychology) Speech recognition Mathematics Linguistics

Metrics

15
Cited By
3.83
FWCI (Field Weighted Citation Impact)
50
Refs
0.92
Citation Normalized Percentile
Is in top 1%
Is in top 10%

Citation History

Topics

Topic Modeling
Physical Sciences →  Computer Science →  Artificial Intelligence
Natural Language Processing Techniques
Physical Sciences →  Computer Science →  Artificial Intelligence
Speech Recognition and Synthesis
Physical Sciences →  Computer Science →  Artificial Intelligence

Related Documents

JOURNAL ARTICLE

DialogueCSE: Dialogue-based Contrastive Learning of Sentence Embeddings

Che LiuRui WangJinghua LiuJian SunFei HuangLuo Si

Journal:   Proceedings of the 2021 Conference on Empirical Methods in Natural Language Processing Year: 2021
JOURNAL ARTICLE

DiffCSE: Difference-based Contrastive Learning for Sentence Embeddings

Yung-Sung ChuangRumen DangovskiHongyin LuoYang ZhangShiyu ChangMarin SoljačićShang-Wen LiScott YihYoon KimJames Glass

Journal:   Proceedings of the 2022 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies Year: 2022 Pages: 4207-4218
© 2026 ScienceGate Book Chapters — All rights reserved.