DiffCSE: Difference between revisions
Jump to navigation
Jump to search
RobowaifuDev (talk | contribs) (Created page with "'''DiffCSE''' is a contrastive learning framework for learning sentence embeddings from unlabeled data.<ref>Chuang et al. "DiffCSE: Difference-based Contrastive Learning for Sentence Embeddings." 2022. [https://arxiv.org/abs/2204.10298 arXiv:2204.10298]</ref> It uses a difference-based loss function to compare two sentence embeddings, one generated by masking out the sentence and filling it in with generated data with a masked language model, and training a model to prod...") |
RobowaifuDev (talk | contribs) No edit summary |
||
(One intermediate revision by the same user not shown) | |||
Line 1: | Line 1: | ||
'''DiffCSE''' is a contrastive learning framework for learning sentence embeddings from unlabeled data.<ref>Chuang et al. "DiffCSE: Difference-based Contrastive Learning for Sentence Embeddings." 2022. [https://arxiv.org/abs/2204.10298 arXiv:2204.10298]</ref> It uses a difference-based loss function to compare two sentence embeddings, one generated by masking out the sentence and filling it in with generated data with a masked language model, and training a model to produce representations that accurately capture the semantic similarity between sentences. This approach is demonstrated to be effective for a variety of downstream tasks, such as multi-document summarization, sentence similarity assessment, and text classification. | '''DiffCSE''' is a contrastive learning framework for learning [[Sentence embedding|sentence embeddings]] from unlabeled data.<ref>Chuang et al. "DiffCSE: Difference-based Contrastive Learning for Sentence Embeddings." 2022. [https://arxiv.org/abs/2204.10298 arXiv:2204.10298]</ref> It uses a difference-based loss function to compare two sentence embeddings, one generated by masking out the sentence and filling it in with generated data with a masked language model, and training a model to produce representations that accurately capture the semantic similarity between sentences. This approach is demonstrated to be effective for a variety of downstream tasks, such as multi-document summarization, sentence similarity assessment, and text classification. | ||
A [[RoBERTa]] model finetuned with DiffCSE is available on [https://huggingface.co/voidism/diffcse-roberta-base-sts Hugging Face]. | |||
=== References === |
Latest revision as of 15:12, 28 April 2023
DiffCSE is a contrastive learning framework for learning sentence embeddings from unlabeled data.[1] It uses a difference-based loss function to compare two sentence embeddings, one generated by masking out the sentence and filling it in with generated data with a masked language model, and training a model to produce representations that accurately capture the semantic similarity between sentences. This approach is demonstrated to be effective for a variety of downstream tasks, such as multi-document summarization, sentence similarity assessment, and text classification.
A RoBERTa model finetuned with DiffCSE is available on Hugging Face.
References
- ↑ Chuang et al. "DiffCSE: Difference-based Contrastive Learning for Sentence Embeddings." 2022. arXiv:2204.10298