Skip to content

Unsupervised Contrastive Learning of Sentence Embeddings through Optimized Sample Construction and Knowledge Distillation

Notifications You must be signed in to change notification settings

dingyan0352/dyfinalcode

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 

History

3 Commits
 
 
 
 
 
 

Repository files navigation

dyfinalcode

Unsupervised Contrastive Learning of Sentence Embeddings through Optimized Sample Construction and Knowledge Distillation

Unsupervised contrastive learning of sentence embedding has been a recent focus of researchers. However, issues such as unreasonable division of positive and negative samples and poor data enhancementleading to text semantic changes still exist. We propose an optimized data augmentation method that combines contrastive learning’s data augmentation with unsupervised sentence pair modelling’s distillation. Our data augmentation uses in-sentence tokens for positive examples and text similarity for negative examples, while the distillation is conducted without supervised pairs. Experimental results on the STS task show that our method achieves a Spearman correlation of 81.03%, outperforming existing STS benchmarks.

About

Unsupervised Contrastive Learning of Sentence Embeddings through Optimized Sample Construction and Knowledge Distillation

Resources

Stars

Watchers

Forks

Releases

No releases published

Packages