UNSEE: Unsupervised Non-contrastive Sentence Embeddings

Ömer Çağatan


Abstract
In this paper, we introduce UNSEE, which stands for Unsupervised Non-Contrastive Sentence Embeddings. UNSEE demonstrates better performance compared to SimCSE in the Massive Text Embedding (MTEB) benchmark. We begin by highlighting the issue of representation collapse that occurs with the replacement of contrastive objectives with non-contrastive objectives in SimCSE. Subsequently, we introduce a straightforward solution called the target network to mitigate this problem. This approach enables us to harness non-contrastive objectives while ensuring training stability and achieving performance improvements similar to those seen with contrastive objectives. We have reached peak performance in non-contrastive sentence embeddings through extensive fine-tuning and optimization. These efforts have resulted in superior sentence representation models, emphasizing the importance of careful tuning and optimization for non-contrastive objectives.
Anthology ID:
2024.eacl-long.23
Volume:
Proceedings of the 18th Conference of the European Chapter of the Association for Computational Linguistics (Volume 1: Long Papers)
Month:
March
Year:
2024
Address:
St. Julian’s, Malta
Editors:
Yvette Graham, Matthew Purver
Venue:
EACL
SIG:
Publisher:
Association for Computational Linguistics
Note:
Pages:
384–393
Language:
URL:
https://aclanthology.org/2024.eacl-long.23
DOI:
Bibkey:
Cite (ACL):
Ömer Çağatan. 2024. UNSEE: Unsupervised Non-contrastive Sentence Embeddings. In Proceedings of the 18th Conference of the European Chapter of the Association for Computational Linguistics (Volume 1: Long Papers), pages 384–393, St. Julian’s, Malta. Association for Computational Linguistics.
Cite (Informal):
UNSEE: Unsupervised Non-contrastive Sentence Embeddings (Çağatan, EACL 2024)
Copy Citation:
PDF:
https://aclanthology.org/2024.eacl-long.23.pdf