Skip to content

xbdxwyh/mocose

Repository files navigation

MoCoSE

Implementation of paper: Exploring the Impact of Negative Samples of Contrastive Learning: A Case Study of Sentence Embedding , been accepted to appear in the Findings of ACL 2022.

We propose a momentum contrastive learning model to sentence embedding, namely MoCoSE. We focus on the effect of negative queue length in text comparison learning.

Attention! You may need to:

  1. download the BERT weights and change the path of the weights in demo.
  2. download the sentEval and change the corresponding path in mocose_tools.

Model Structure

architecture

  • mocose.py contains the main constituent code of the model;
  • mocose_tools.py contains the code of the tools to evaluate the model;
  • mocose_demo.ipynbs is the example code we provide for train and evaluation.

You can download MoCoSE-bert-base-uncased weights HERE .

STS Results in our paper:

STS12 STS13 STS14 STS15 STS16 STS-Benchmark SICK-R Avg.
71.48 81.40 74.47 83.45 78.99 78.68 72.44 77.27

Requirement

  • pytorch 1.9.0
  • typing 4.0.1
  • transformers 4.11.3
  • datasets 1.5.0
  • nlpaug 1.1.10
  • tqdm 4.49.0
  • PrettyTable 2.1.0

About

No description, website, or topics provided.

Resources

License

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published