princeton-nlp / SimCSE

[EMNLP 2021] SimCSE: Simple Contrastive Learning of Sentence Embeddings https://arxiv.org/abs/2104.08821

Geek Repo:Geek Repo

Github PK Tool:Github PK Tool

[question] Pretrined sentence embeddings model fine tuning

nurulchamidah opened this issue · comments

Hello, thanks for your great and inspiring works .
I would like to ask, is fine tuning in supervised manner of sentence embeddings that is produced using unsupervised learning or other way round possible?
OR pretrained sentence embeddings from model produced from supervised triplets (with hard negatives), fine tuned using non triplet (example : entailment pair) in another dataset. is this possible?
thanks before.

Hi, thanks for your attention. That's an interesting idea, and I believe it's possible. For your reference, a recent work BGE is first tuned with large-scale automatically extracted sequence pairs and then tuned on high-quality datasets like NLI datasets.

oke. thank you for your answer and reference. it helps me a lot.