TY - JOUR
T1 - Contrastive sentence representation learning with adaptive false negative cancellation
AU - Xu, Lingling
AU - Xie, Haoran
AU - Wang, Fu Lee
AU - Tao, Xiaohui
AU - Wang, Weiming
AU - Li, Qing
N1 - Publisher Copyright:
© 2023 The Authors
PY - 2024/2
Y1 - 2024/2
N2 - Contrastive sentence representation learning has made great progress thanks to a range of text augmentation strategies and hard negative sampling techniques. However, most studies directly employ in-batch samples as negative samples, ignoring the semantic relationship between negative samples and anchors, which may lead to negative sampling bias. To address this issue, we propose similarity and relative-similarity strategies for identifying potential false negatives. Moreover, we introduce adaptive false negative elimination and attraction methods to mitigate their adverse effects. Our proposed approaches can also be considered semi-supervised contrastive learning, as the identified false negatives can be viewed as either negative or positive samples for contrastive learning in adaptive false negative elimination and attraction methods. By fusing information from positive and negative pairs, contrastive learning learns rich and discriminative representations that capture the intrinsic characteristics of the sentence. Experimental results indicate that our proposed strategies and methods can bring further significant performance improvements. Specifically, the combination of similarity strategy and adaptive false negative elimination method achieves the best results, yielding an average performance gain of 2.1% compared to SimCSE in semantic textual similarity (STS) tasks. Furthermore, our approach is generalizable and can be applied to different text data augmentation strategies and certain existing contrastive sentence representation learning models. Our experimental code and data are publicly available at the link: https://github.com/Linda230/AFNC.
AB - Contrastive sentence representation learning has made great progress thanks to a range of text augmentation strategies and hard negative sampling techniques. However, most studies directly employ in-batch samples as negative samples, ignoring the semantic relationship between negative samples and anchors, which may lead to negative sampling bias. To address this issue, we propose similarity and relative-similarity strategies for identifying potential false negatives. Moreover, we introduce adaptive false negative elimination and attraction methods to mitigate their adverse effects. Our proposed approaches can also be considered semi-supervised contrastive learning, as the identified false negatives can be viewed as either negative or positive samples for contrastive learning in adaptive false negative elimination and attraction methods. By fusing information from positive and negative pairs, contrastive learning learns rich and discriminative representations that capture the intrinsic characteristics of the sentence. Experimental results indicate that our proposed strategies and methods can bring further significant performance improvements. Specifically, the combination of similarity strategy and adaptive false negative elimination method achieves the best results, yielding an average performance gain of 2.1% compared to SimCSE in semantic textual similarity (STS) tasks. Furthermore, our approach is generalizable and can be applied to different text data augmentation strategies and certain existing contrastive sentence representation learning models. Our experimental code and data are publicly available at the link: https://github.com/Linda230/AFNC.
KW - Adaptive weight
KW - Contrastive learning
KW - Negative sampling bias
KW - Semi-supervised learning
KW - Sentence representation learning
UR - http://www.scopus.com/inward/record.url?scp=85174687747&partnerID=8YFLogxK
U2 - 10.1016/j.inffus.2023.102065
DO - 10.1016/j.inffus.2023.102065
M3 - Article
AN - SCOPUS:85174687747
SN - 1566-2535
VL - 102
JO - Information Fusion
JF - Information Fusion
M1 - 102065
ER -