Awesome-Sentence-Embedding
A curated list of research papers in Sentence Reprsentation Learning.
The leaderboard of Unsupervised STS is also available.
Unsupervised STS leaderboard
You can search by model name (alias) to find the corresponding paper in the Related Papers.
Methods |
STS12 |
STS13 |
STS14 |
STS15 |
STS16 |
STS-B |
SICK-R |
avg. |
GloVe(avg.) |
55.14 |
70.66 |
59.73 |
68.25 |
63.66 |
58.02 |
53.76 |
61.32 |
USE |
64.49 |
67.8 |
64.61 |
76.83 |
73.18 |
74.92 |
76.69 |
71.22 |
CLS |
21.54 |
32.11 |
21.28 |
37.89 |
44.24 |
20.3 |
42.42 |
31.4 |
Mean |
30.87 |
59.89 |
47.73 |
60.29 |
63.73 |
47.29 |
58.22 |
52.57 |
first-last avg. |
39.7 |
59.38 |
49.67 |
66.03 |
66.19 |
53.87 |
62.06 |
56.7 |
Contrastive(BT) |
54.26 |
64.03 |
54.28 |
68.19 |
67.5 |
63.27 |
66.91 |
62.63 |
BERT-Flow |
58.4 |
67.1 |
60.85 |
75.16 |
71.22 |
68.66 |
64.47 |
66.55 |
BERT-Whitening |
57.83 |
66.9 |
60.9 |
75.08 |
71.31 |
68.24 |
63.73 |
66.28 |
IS-BERT |
56.77 |
69.24 |
61.21 |
75.23 |
70.16 |
69.21 |
64.25 |
66.58 |
BSL |
67.83 |
71.4 |
66.88 |
79.97 |
73.97 |
73.74 |
70.4 |
72.03 |
CT-BERT |
61.63 |
76.8 |
68.47 |
77.5 |
76.48 |
74.31 |
69.19 |
72.05 |
ConSERT |
64.64 |
78.49 |
69.07 |
79.72 |
75.95 |
73.97 |
67.31 |
72.74 |
SCD |
66.94 |
78.03 |
69.89 |
78.73 |
76.23 |
76.3 |
73.18 |
74.19 |
SG-OPT |
66.84 |
80.13 |
71.23 |
81.56 |
77.17 |
77.23 |
68.16 |
74.62 |
Mirror-BERT |
69.1 |
81.1 |
73 |
81.9 |
75.7 |
78 |
69.1 |
75.4 |
PaSeR |
70.21 |
83.88 |
73.06 |
83.87 |
77.60 |
79.19 |
65.31 |
76.16 |
SimCSE |
68.4 |
82.41 |
74.38 |
80.91 |
78.56 |
76.85 |
72.23 |
76.25 |
MCSE+coco |
71.2±1.3 |
79.7±0.9 |
73.8±0.9 |
83.0±0.4 |
77.8±0.9 |
78.5±0.4 |
72.1±1.4 |
76.6±0.5 |
EASE |
72.8 |
81.8 |
73.7 |
82.3 |
79.5 |
78.9 |
69.7 |
77 |
L2P-CSR |
70.21 |
83.25 |
75.42 |
82.34 |
78.75 |
77.8 |
72.65 |
77.2 |
DCLR |
70.81 |
83.73 |
75.11 |
82.56 |
78.44 |
78.31 |
71.59 |
77.22 |
MoCoSE |
71.58 |
81.4 |
74.47 |
83.45 |
78.99 |
78.68 |
72.44 |
77.27 |
InforMin-CL |
70.22 |
83.48 |
75.51 |
81.72 |
79.88 |
79.27 |
71.03 |
77.3 |
MCSE+flickr |
71.4±0.9 |
81.8±1.3 |
74.8±0.9 |
83.6±0.9 |
77.5±0.8 |
79.5±0.5 |
72.6±1.4 |
77.3±0.5 |
VisualCSE |
71.16 |
83.29 |
75.13 |
81.59 |
80.05 |
80.03 |
71.23 |
77.5 |
SimCSE+GS-InfoNCE |
70.12 |
82.57 |
75.21 |
82.89 |
80.23 |
79.7 |
72.7 |
77.63 |
MixCSE |
71.71±4.04 |
83.14±0.72 |
75.49±1.25 |
83.64±2.32 |
79.00±0.16 |
78.48±0.82 |
72.19±0.46 |
77.66±0.61 |
PT-BERT |
71.2 |
83.76 |
76.34 |
82.63 |
78.9 |
79.42 |
71.94 |
77.74 |
AudioCSE |
71.65 |
84.27 |
76.69 |
83.22 |
78.69 |
79.94 |
70.49 |
77.85 |
ArcCSE |
72.08 |
84.27 |
76.25 |
82.32 |
79.54 |
79.92 |
72.39 |
78.11 |
miCSE |
71.71 |
83.09 |
75.46 |
83.13 |
80.22 |
79.7 |
73.62 |
78.13 |
PCL |
72.74 |
83.36 |
76.05 |
83.07 |
79.26 |
79.72 |
72.75 |
78.14 |
ESimCSE |
73.4 |
83.27 |
77.25 |
82.66 |
78.81 |
80.17 |
72.3 |
78.27 |
DiffCSE |
72.28 |
84.43 |
76.47 |
83.9 |
80.54 |
80.59 |
71.23 |
78.49 |
PromptBERT |
71.56 |
84.58 |
76.98 |
84.47 |
80.6 |
81.6 |
69.87 |
78.54 |
InfoCSE |
70.53 |
84.59 |
76.40 |
85.10 |
81.95 |
82.00 |
71.37 |
78.85 |
SNCSE |
70.67 |
84.79 |
76.99 |
83.69 |
80.51 |
81.35 |
74.77 |
78.97 |
Prompt+L2P-CSR |
72.34 |
84.81 |
78.13 |
84.16 |
80.58 |
82.04 |
71.13 |
79.03 |
RankCSE_listNet |
74.38 |
85.97 |
77.51 |
84.46 |
81.31 |
81.46 |
75.26 |
80.05 |
RankEncoder |
74.88 |
85.59 |
78.61 |
83.5 |
80.56 |
81.55 |
75.78 |
80.07 |
RankCSE_listMLE |
75.66 |
86.27 |
77.81 |
84.74 |
81.1 |
81.8 |
75.13 |
80.36 |
Methods |
STS12 |
STS13 |
STS14 |
STS15 |
STS16 |
STS-B |
SICK-R |
avg. |
CLS |
27.44 |
30.76 |
22.59 |
29.98 |
42.74 |
26.75 |
43.44 |
31.96 |
Mean |
27.67 |
55.79 |
44.49 |
51.67 |
61.88 |
47 |
53.85 |
48.9 |
Contrastive(BT) |
52.04 |
62.59 |
54.25 |
71.07 |
66.71 |
63.84 |
66.53 |
62.43 |
BERT-Flow |
62.82 |
71.24 |
65.39 |
78.98 |
73.23 |
72.72 |
63.77 |
70.07 |
BERT-Whitening |
64.34 |
74.6 |
69.64 |
74.68 |
75.9 |
72.48 |
60.8 |
70.35 |
SG-OPT |
67.02 |
79.42 |
70.38 |
81.72 |
76.35 |
76.16 |
70.2 |
74.46 |
ConSERT |
70.69 |
82.96 |
74.13 |
82.78 |
76.66 |
77.53 |
70.37 |
76.45 |
SimCSE |
70.88 |
84.16 |
76.43 |
84.5 |
79.76 |
79.26 |
73.88 |
78.41 |
MixCSE |
72.55±0.49 |
84.32±0.53 |
76.69±0.76 |
84.31±0.10 |
79.67±0.28 |
79.90±0.18 |
74.07±0.13 |
78.80±0.09 |
DCLR |
71.87 |
84.83 |
77.37 |
84.7 |
79.81 |
79.55 |
74.19 |
78.9 |
L2P-CSR |
71.44 |
85.09 |
76.88 |
84.71 |
80 |
79.75 |
74.55 |
78.92 |
SimCSE+GS-InfoNCE |
73.75 |
85.09 |
77.35 |
84.44 |
79.88 |
79.94 |
73.48 |
78.96 |
MoCoSE |
74.5 |
84.54 |
77.32 |
84.11 |
79.67 |
80.53 |
73.26 |
79.13 |
ESimCSE |
73.21 |
85.37 |
77.73 |
84.3 |
78.92 |
80.73 |
74.89 |
79.31 |
ArcCSE |
73.17 |
86.19 |
77.9 |
84.97 |
79.43 |
80.45 |
73.5 |
79.37 |
PromptBERT |
73.29 |
86.39 |
77.9 |
85.18 |
79.97 |
81.92 |
71.26 |
79.42 |
Prompt+L2P-CSR |
73.14 |
86.78 |
78.67 |
85.77 |
80.32 |
82.23 |
72.57 |
79.93 |
PCL |
74.89 |
85.88 |
78.33 |
85.3 |
80.13 |
81.39 |
73.66 |
79.94 |
InfoCSE |
71.89 |
86.17 |
77.72 |
86.20 |
81.29 |
83.16 |
74.84 |
80.18 |
SNCSE |
71.94 |
86.66 |
78.84 |
85.74 |
80.72 |
82.29 |
75.11 |
80.19 |
RankCSE_listNet |
74.75 |
86.46 |
78.52 |
85.41 |
80.62 |
81.4 |
76.12 |
80.47 |
RankCSE_listMLE |
75.48 |
86.5 |
78.6 |
85.45 |
81.09 |
81.58 |
75.53 |
80.6 |
Methods |
STS12 |
STS13 |
STS14 |
STS15 |
STS16 |
STS-B |
SICK-R |
avg. |
CLS |
16.67 |
45.57 |
30.36 |
55.08 |
56.98 |
45.41 |
61.89 |
44.57 |
Mean |
32.11 |
56.33 |
45.22 |
61.34 |
61.98 |
54.53 |
62.03 |
53.36 |
BERT-Whitening |
46.99 |
63.24 |
57.23 |
71.36 |
68.99 |
61.36 |
62.91 |
61.73 |
DeCLUTR |
52.41 |
75.19 |
65.52 |
77.12 |
78.63 |
72.41 |
68.62 |
69.99 |
BSL |
68.47 |
72.41 |
68.48 |
78.5 |
72.77 |
78.77 |
69.97 |
72.76 |
SG-OPT |
62.57 |
78.96 |
69.24 |
79.99 |
77.17 |
77.6 |
68.42 |
73.42 |
SCD |
63.53 |
77.79 |
69.79 |
80.21 |
77.29 |
76.55 |
72.1 |
73.89 |
Contrastive(BT) |
62.34 |
78.6 |
68.65 |
79.31 |
77.49 |
79.93 |
71.97 |
74.04 |
Mirror-BERT |
66.6 |
82.7 |
74 |
82.4 |
79.7 |
79.6 |
69.7 |
76.4 |
SimCSE |
70.16 |
81.77 |
73.24 |
81.36 |
80.65 |
80.22 |
68.56 |
76.57 |
EASE |
70.9 |
81.5 |
73.5 |
82.6 |
80.5 |
80 |
68.4 |
76.8 |
VaSCL |
69.02 |
82.38 |
73.93 |
82.54 |
80.96 |
69.4 |
80.52 |
76.96 |
InforMin-CL |
69.79 |
82.57 |
73.36 |
80.91 |
81.28 |
81.07 |
70.3 |
77.04 |
ESimCSE |
69.9 |
82.5 |
74.68 |
83.19 |
80.3 |
80.99 |
70.54 |
77.44 |
MCSE+coco |
70.2±1.7 |
82.0±0.7 |
75.5±1.2 |
83.0±0.6 |
81.5±0.7 |
80.8±1.0 |
69.9±0.6 |
77.6±0.8 |
SimCSE+GS-InfoNCE |
71.12 |
83.24 |
75 |
82.61 |
81.36 |
81.26 |
69.62 |
77.74 |
L2P-CSR |
71.69 |
82.43 |
74.55 |
82.15 |
81.81 |
81.36 |
70.22 |
77.74 |
AudioCSE |
68.44 |
83.96 |
75.77 |
82.38 |
82.07 |
81.63 |
70.56 |
77.83 |
DCLR |
70.01 |
83.08 |
75.09 |
83.66 |
81.06 |
81.86 |
70.33 |
77.87 |
VisualCSE |
70.41 |
83.51 |
74.87 |
82.79 |
81.67 |
81.89 |
69.95 |
77.87 |
PCL |
71.54 |
82.7 |
75.38 |
83.31 |
81.64 |
81.61 |
69.19 |
77.91 |
DiffCSE |
70.05 |
83.43 |
75.49 |
82.81 |
82.12 |
82.38 |
71.19 |
78.21 |
MCSE+flickr |
71.7±0.2 |
82.7±0.4 |
75.9±0.3 |
84.0±0.4 |
81.3±0.3 |
82.3±0.5 |
70.3±1.3 |
78.3±0.1 |
CARDS |
72.49 |
84.09 |
76.19 |
82.98 |
82.11 |
82.25 |
70.65 |
78.68 |
PromptBERT |
73.94 |
84.74 |
77.28 |
84.99 |
81.74 |
81.88 |
69.5 |
79.15 |
SNCSE |
70.62 |
84.42 |
77.24 |
84.85 |
81.49 |
83.07 |
72.92 |
79.23 |
RankCSE_listMLE |
72.74 |
84.24 |
75.99 |
84.68 |
82.88 |
83.16 |
71.77 |
79.35 |
RankCSE_listNet |
72.88 |
84.5 |
76.46 |
84.67 |
83 |
83.24 |
71.67 |
79.49 |
Prompt+L2P-CSR |
74.97 |
83.63 |
78.28 |
84.86 |
82.03 |
82.77 |
71.26 |
79.69 |
Methods |
STS12 |
STS13 |
STS14 |
STS15 |
STS16 |
STS-B |
SICK-R |
avg. |
CLS |
19.25 |
22.97 |
14.93 |
33.41 |
38.01 |
12.52 |
40.63 |
25.96 |
Mean |
33.63 |
57.22 |
45.67 |
63 |
61.18 |
47.07 |
58.38 |
52.31 |
Contrastive(BT) |
57.6 |
72.14 |
62.25 |
71.49 |
71.75 |
77.05 |
67.83 |
68.59 |
BERT-Whitening |
64.17 |
73.92 |
71.06 |
76.4 |
74.87 |
71.68 |
58.49 |
70.08 |
SG-OPT |
64.29 |
76.36 |
68.48 |
80.1 |
76.6 |
78.14 |
67.97 |
73.13 |
InforMin-CL |
70.91 |
84.2 |
75.57 |
82.26 |
79.68 |
81.1 |
72.81 |
78.08 |
SimCSE |
72.86 |
83.99 |
75.62 |
84.77 |
81.8 |
81.98 |
71.26 |
78.9 |
VaSCL |
73.36 |
83.55 |
77.16 |
83.25 |
80.66 |
72.96 |
82.36 |
79.04 |
SimCSE+GS-InfoNCE |
71.76 |
84.91 |
76.79 |
84.35 |
81.74 |
82.97 |
71.71 |
79.21 |
DCLR |
73.09 |
84.57 |
76.13 |
85.15 |
81.99 |
82.35 |
71.8 |
79.3 |
PCL |
73.76 |
84.59 |
76.81 |
85.37 |
81.66 |
82.89 |
70.33 |
79.34 |
PromptBERT |
73.24 |
83.08 |
77.97 |
84.03 |
81.57 |
82.85 |
73.28 |
79.43 |
ESimCSE |
73.2 |
84.93 |
76.88 |
84.86 |
81.21 |
82.79 |
72.27 |
79.45 |
AudioCSE |
72.1 |
84.3 |
76.74 |
85.11 |
82.51 |
82.94 |
72.45 |
79.45 |
L2P-CSR |
73.29 |
84.08 |
76.65 |
85.47 |
82.7 |
82.15 |
72.36 |
79.53 |
VisualCSE |
73.09 |
84.77 |
77.09 |
85.47 |
82.06 |
83.26 |
72.23 |
79.71 |
Prompt+L2P-CSR |
73.65 |
84.08 |
78.29 |
85.36 |
82.15 |
83.7 |
73.47 |
80.1 |
RankCSE_listMLE |
73.4 |
85.34 |
77.25 |
85.45 |
82.64 |
84.14 |
72.92 |
80.16 |
RankCSE_listNet |
73.23 |
85.08 |
77.5 |
85.67 |
82.99 |
84.2 |
72.98 |
80.24 |
CARDS |
74.63 |
86.27 |
79.25 |
85.93 |
83.17 |
83.86 |
72.77 |
80.84 |
SNCSE |
73.71 |
86.73 |
80.35 |
86.8 |
83.06 |
84.31 |
77.43 |
81.77 |
Main Track of Sentence Embeddings
Tips:
- 【Conf】Titile【Alias of the paper, etc.】
- 【ICLR2023 withdrawn】 RankCSE: Unsupervised Sentence Representations Learning via Learning to Rank 【RankCSE】paper
- 【ICLR2023 withdrawn】 Ranking-Enhanced Unsupervised Sentence Representation Learning 【RankEncoder】paper
- 【ICLR2023 withdrawn】 Learning to Perturb for Contrastive Learning of Unsupervised Sentence Representations 【L2P-CSR】paper
- 【ICLR2023 withdrawn】 miCSE: Mutual Information Contrastive Learning for Low-shot Sentence Embeddings 【miCSE】
- 【ICLR2023】On The Inadequacy of Optimizing Alignment and Uniformity in Contrastive Learning of Sentence Representations
- 【EMNLP2022】ConGen: Unsupervised Control and Generalization Distillation For Sentence Representation
- 【EMNLP2022】Generate, Discriminate and Contrast: A Semi-Supervised Sentence Representation Learning Framework
- 【EMNLP2022】 PCL: Peer-Contrastive Learning with Diverse Augmentations for Unsupervised Sentence Embeddings 【PCL】
- 【EMNLP2022】Contrastive Learning with Prompt-derived Virtual Semantic Prototypes for Unsupervised Sentence Embedding 【PromptCSE】
- 【EMNLP2022】Sentence Representation Learning with Generative Objective rather than Contrastive Objective 【PaSeR, Unsupervised & Supervised STS】
- 【EMNLP2022】DiffAug: Differential Data Augmentation for Contrastive Sentence Representation Learning 【DiffAug, Semi-supervised, Supervised STS】
- 【EMNLP2022】InfoCSE: Information-aggregated Contrastive Learning of Sentence Embeddings 【InfoCSE】
- 【EMNLP2022】 Improved Universal Sentence Embeddings with Prompt-based Contrastive Learning and Energy-based Learning 【PromCSE, Supervised STS】
- 【EMNLP2022】 PromptBERT: Improving BERT Sentence Embeddings with Prompts 【PromptBERT】
- 【COLING2022】 Smoothed contrastive learning for unsupervised sentence embedding 【GS-InfoNCE】
- 【COLING2022】 ESimCSE: Enhanced Sample Building Method for Contrastive Learning of Unsupervised Sentence Embedding 【ESimCSE】
- 【COLING2022】 An information minimization contrastive learning model for unsupervised sentence embeddings learning 【InforMin-CL】
- 【NIPS2022】 Non-Linguistic Supervision for Contrastive Learning of Sentence Embeddings 【VisualCSE, AudioCSE, more data】
- 【SIGIR2022】 Improving Contrastive Learning of Sentence Embeddings with Case-Augmented Positives and Retrieved Negatives 【CARDS】
- 【AAAI2022】 Unsupervised Sentence Representation via Contrastive Learning with Mixing Negatives 【MixCSE】
- 【ACL2022】 A contrastive framework for learning sentence representations from pairwise and triple-wise perspective in angular space 【ArcCSE】
- 【ACL2022】 Debiased contrastive learning of unsupervised sentence representations 【DCLR】
- 【ACL2022】 Virtual augmentation supported contrastive learning of sentence representations 【VaSCL】
- 【ACL2022】 Scd: Self-contrastive decorrelation for sentence embeddings 【SCD】
- 【ACL2022】 Exploring the impact of negative samples of contrastive learning: A case study of sentence embedding 【MoCoSE】
- 【ACL2022】 A Sentence is Worth 128 Pseudo Tokens: A Semantic-Aware Contrastive Learning Framework for Sentence Embeddings 【PT-BERT】
- 【ACL2022】 Sentence-T5: Scalable Sentence Encoders from Pre-trained Text-to-Text Models 【ST5, Supervised STS】
- 【NAACL2022】 DiffCSE:Difference-based Contrastive Learning for Sentence Embeddings 【DiffCSE】
- 【NAACL2022】 EASE: Entity-Aware Contrastive Learning of Sentence Embedding 【EASE】
- 【NAACL2022】 MCSE: Multimodal Contrastive Learning of Sentence Embeddings 【MCSE, more data】
- 【ICML2022ws】 Boosting Monolingual Sentence Representation with Large-scale Parallel Translation Datasets 【BMSR, Supervised STS, more data】
- 【Arxiv2022】 SNCSE: Contrastive Learning for Unsupervised Sentence Embedding with Soft Negative Samples 【SNCSE】
- 【ICLR2021】 Trans-encoder: Unsupervised sentence-pair modelling through self-and mutual-distillations 【TENC, use unlabeled STS data】
- 【ICLR2021】 Semantic re-tuning with contrastive tension 【CT-BERT】
- 【Arxiv2021】 Whitening sentence representations for better semantics and faster retrieval 【BERT-whitening】
- 【Arxiv2021】 DisCo: Effective Knowledge Distillation For Contrastive Learning of Sentence Embeddings 【DisCo, Supervised STS】
- 【EMNLP2021】 Locality Preserving Sentence Encoding 【SBERT-LP, Supervised STS】
- 【EMNLP2021】 Universal Sentence Representation Learning with Conditional Masked Language Model 【CMLM, Supervised STS】
- 【EMNLP2021】 PAUSE: Positive and Annealed Unlabeled Sentence Embedding 【PAUSE, Supervised STS】
- 【EMNLP2021】 SimCSE: Simple contrastive learning of sentence embeddings 【SimCSE】
- 【EMNLP2021】 Fast, effective, and self-supervised: Transforming masked language models into universal lexical and sentence encoders 【Mirror-BERT】
- 【EMNLP2021】 Pairwise Supervised Contrastive Learning of Sentence Representations 【PairSupCon, Supervised STS】
- 【NAACL2021】 Disentangling Semantics and Syntax in Sentence Embeddings with Pre-trained Language Models 【ParaBART, Supervised STS】
- 【ACL2021】 ConSERT: A contrastive framework for self-supervised sentence representation transfer 【ConSERT, use unlabeled STS data】
- 【ACL2021】 DeCLUTR: Deep contrastive learning for unsupervised textual representations 【DeCLUTR】
- 【ACL2021】 Bootstrapped unsupervised sentence representation learning 【BSL, Unsup & Sup】
- 【ACL2021】 Self-guided contrastive learning for bert sentence representations 【SG-OPT】
- 【SIGIR2021】 Dual-View Distilled BERT for Sentence Embedding 【DvBERT, Supervised STS】
- 【EMNLP2020】 On the sentence embeddings from pre-trained language models 【BERT-flow】
- 【EMNLP2020】 An unsupervised sentence embedding method by mutual information maximization 【IS-BERT】
- 【TASLP2020】 SBERT-WK: A Sentence Embedding Method by Dissecting BERT-Based Word Models 【SBERT-WK, Supervised STS】
- 【EMNLP2019】 Sentence-bert: Sentence embeddings using siamese bert-networks 【SBERT】
- 【ICLR2018】 An efficient framework for learning sentence representations 【Quick-thought】
- 【EMNLP2018】 Universal Sentence Encoder for English 【USE】
- 【EMNLP2017】 Supervised learning of universal sentence representations from natural language inference data 【InferSent】
- 【ACL2016】 Learning distributed representations of sentences from unlabelled data 【FastSent】
- 【NIPS2015】 Skip-thought vectors 【Skip-thought】
Others Track of Sentence Embeddings
- 【ACL2022】 Compressing Sentence Representation for Semantic Retrieval via Homomorphic Projective Distillation
- 【NAACL2022】 Learning Dialogue Representations from Consecutive Utterances
- 【NAACL2022】 On the Effectiveness of Sentence Encoding for Intent Detection Meta-Learning
- 【ICASSP2022】 Integrating Dependency Tree into Self-Attention for Sentence Representation
- 【ICASSP2022】 Pair-Level Supervised Contrastive Learning for Natural Language Inference
- 【Arxiv2022】 Masked Autoencoders As The Unified Learners For Pre-Trained Sentence Representation
- 【Arxiv2022】 RetroMAE: Pre-training Retrieval-oriented Transformers via Masked Auto-Encoder
- 【EMNLP2021】 A Deep Decomposable Model for Disentangling Syntax and Semantics in Sentence Representation
- 【EMNLP2021】 DialogueCSE: Dialogue-based Contrastive Learning of Sentence Embeddings
- 【EMNLP2021】 WhiteningBERT: An Easy Unsupervised Sentence Embedding Approach
- 【EMNLP2021】 TSDAE: Using Transformer-based Sequential Denoising Auto-Encoderfor Unsupervised Sentence Embedding Learning
- 【EMNLP2021】 Exploiting Twitter as Source of Large Corpora of Weakly Similar Pairs for Semantic Sentence Embeddings
- 【ACL2021】 DefSent: Sentence Embeddings using Definition Sentences
- 【ACL2021】 Discrete Cosine Transform as Universal Sentence Encoder
- 【Arxiv2021】 S-SimCSE: sampled sub-networks for contrastive learning of sentence embedding
- 【NAACL2021】 Augmented SBERT: Data augmentation method for improving bi-encoders for pairwise sentence scoring tasks
- 【Arxiv2020】 CERT: contrastive self-supervised learning for language understanding
- 【AAAI2020】 Scalable Attentive Sentence-Pair Modeling via Distilled Sentence Embedding
- 【EMNLP2020】 Cross-Thought for Sentence Encoder Pre-training
- 【WWW2020】 Enhanced-RCNN: An Efficient Method for Learning Sentence Similarity
- 【ACL2020】 QuASE: Question-Answer Driven Sentence Encoding
- 【EMNLP2019】 Efficient Sentence Embedding using Discrete Cosine Transform
- 【EMNLP2019】 Parameter-free Sentence Embedding via Orthogonal Basis
- 【EMNLP2019】 Incorporating Visual Semantics into Sentence Representations within a Grounded Space
- 【NAACL2019】 Continual Learning for Sentence Representations Using Conceptors
- 【NAACL2019】 A Multi-Task Approach for Disentangling Syntax and Semantics in Sentence Representations
- 【ACL2019】 Simple and Effective Paraphrastic Similarity from Parallel Translations
- 【ACL2019】 Learning Compressed Sentence Representations for On-Device Text Processing
- 【ACL2019】 Exploiting Invertible Decoders for Unsupervised Sentence Representation Learning
Cross-lingual Sentence Embeddings
- 【EMNLP2022】A Multilingual Generative Transformer for Semantic Sentence Embedding
- 【EMNLP2022】English Contrastive Learning Can Learn Universal Cross-lingual Sentence Embeddings 【mSimCSE】
- 【ACL2022】 Language-agnostic BERT Sentence Embedding
- 【IJCAI2022】 Unsupervised Context Aware Sentence Representation Pretraining for Multi-lingual Dense Retrieval
- 【Arxiv2021】 Paraphrastic Representations at Scale
- 【ACL2021】 Lightweight Cross-Lingual Sentence Representation Learning
- 【ACL2021】 Bootstrapped Unsupervised Sentence Representation Learning
- 【ACL2021】 Self-Guided Contrastive Learning for BERT Sentence Representations
- 【EMNLP2021】 Fast, Effective and Self-Supervised: Transforming Masked Language Models into Universal Lexical and Sentence Encoders
- 【EMNLP2021】 Aligning Cross-lingual Sentence Representations with Dual Momentum Contrast
- 【EMNLP2021】 Cross-lingual Sentence Embedding using Multi-Task Learning
- 【EMNLP2021】 Semantic Alignment with Calibrated Similarity for Multilingual Sentence Embedding
- 【EMNLP2021】 Language-agnostic Representation from Multilingual Sentence Encoders for Cross-lingual Similarity Estimation
- 【Arxiv2021】 Analyzing Zero-shot Cross-lingual Transfer in Supervised NLP Tasks
- 【ACL2020】 Unsupervised Multilingual Sentence Embeddings for Parallel Corpus Mining
- 【AAAI2020】 Emu: Enhancing Multilingual Sentence Embeddings with Semantic Specialization
- 【AAAI2020】 Unsupervised Interlingual Semantic Representations from Sentence Embeddings for Zero-Shot Cross-Lingual Transfer
- 【AAAI2020】 ABSent: Cross-Lingual Sentence Representation Mapping with Bidirectional GANs
- 【EMNLP2020】 Making Monolingual Sentence Embeddings Multilingual using Knowledge Distillation
- 【EMNLP2020】 A Bilingual Generative Transformer for Semantic Sentence Embedding
- 【EMNLP2019】 Margin-based Parallel Corpus Mining with Multilingual Sentence Embeddings
- 【EMNLP2019】 Exploring Multilingual Syntactic Sentence Representations
- 【IJCAI2019】 Improving Multilingual Sentence Embedding using Bi-directional Dual Encoder with Additive Margin Softmax
Cross-lingual Dense Retrieval
- 【ICLR2023】 Modeling Sequential Sentence Relation to Improve Cross-lingual Dense Retrieval
- 【ICLR2023】 LEXA: Language-agnostic Cross-consistency Training for Question Answering Tasks
- 【EMNLP2022】Retrofitting Multilingual Sentence Embeddings with Abstract Meaning Representation
- 【IJCAI2022】 Unsupervised Context Aware Sentence Representation Pretraining for Multi-lingual Dense Retrieval
- 【NAACL2022】 EASE: Entity-Aware Contrastive Learning of Sentence Embedding
- 【EMNLP2021】 A Simple and Effective Method To Eliminate the Self Language Bias in Multilingual Representations
- 【EMNLP2020】 Language-Agnostic Answer Retrieval from a Multilingual Pool
- 【ICLR2023】 Understanding The Role of Positional Encodings in Sentence Representations
- 【ICLR2023】 Self-Consistent Learning: Cooperation between Generators and Discriminators
- 【TMLR2022】 Unsupervised dense information retrieval with contrastive learning
- 【NAACL2021】 Supporting Clustering with Contrastive Learning