Dataset Overview

Hint

Quickstart: Find curated datasets or community datasets, choose a loss function via this loss overview, and verify that it works with your dataset.

It is important that your dataset format matches your loss function (or that you choose a loss function that matches your dataset format). See Training Overview > Dataset Format to learn how to verify whether a dataset format works with a loss function.

In practice, most dataset configurations will take one of four forms:

Note that it is often simple to transform a dataset from one format to another, such that it works with your loss function of choice.

Datasets on the Hugging Face Hub

The Datasets library (pip install datasets) allows you to load datasets from the Hugging Face Hub with the load_dataset() function:

from datasets import load_dataset

# Indicate the dataset id from the Hub
dataset_id = "sentence-transformers/natural-questions"
dataset = load_dataset(dataset_id, split="train")
"""
Dataset({
   features: ['query', 'answer'],
   num_rows: 100231
})
"""
print(dataset[0])
"""
{
   'query': 'when did richmond last play in a preliminary final',
   'answer': "Richmond Football Club Richmond began 2017 with 5 straight wins, a feat it had not achieved since 1995. A series of close losses hampered the Tigers throughout the middle of the season, including a 5-point loss to the Western Bulldogs, 2-point loss to Fremantle, and a 3-point loss to the Giants. Richmond ended the season strongly with convincing victories over Fremantle and St Kilda in the final two rounds, elevating the club to 3rd on the ladder. Richmond's first final of the season against the Cats at the MCG attracted a record qualifying final crowd of 95,028; the Tigers won by 51 points. Having advanced to the first preliminary finals for the first time since 2001, Richmond defeated Greater Western Sydney by 36 points in front of a crowd of 94,258 to progress to the Grand Final against Adelaide, their first Grand Final appearance since 1982. The attendance was 100,021, the largest crowd to a grand final since 1986. The Crows led at quarter time and led by as many as 13, but the Tigers took over the game as it progressed and scored seven straight goals at one point. They eventually would win by 48 points – 16.12 (108) to Adelaide's 8.12 (60) – to end their 37-year flag drought.[22] Dustin Martin also became the first player to win a Premiership medal, the Brownlow Medal and the Norm Smith Medal in the same season, while Damien Hardwick was named AFL Coaches Association Coach of the Year. Richmond's jump from 13th to premiers also marked the biggest jump from one AFL season to the next."
}
"""

For more information on how to manipulate your dataset see the Datasets Documentation.

Tip

It’s common for Hugging Face Datasets to contain extraneous columns, e.g. sample_id, metadata, source, type, etc. You can use Dataset.remove_columns to remove these columns, as they will be used as inputs otherwise. You can also use Dataset.select_columns to keep only the desired columns.

Pre-existing Datasets

The Hugging Face Hub hosts 150k+ datasets, many of which can be converted for training embedding models. We are aiming to tag all Hugging Face datasets that work out of the box with Sentence Transformers with sentence-transformers, allowing you to easily find them by browsing to https://huggingface.co/datasets?other=sentence-transformers. We strongly recommend that you browse these datasets to find training datasets that might be useful for your tasks.

These are some of the popular pre-existing datasets tagged as sentence-transformers that can be used to train and fine-tune SentenceTransformer models:

Dataset Description
GooAQ (Question, Answer) pairs from Google auto suggest
Yahoo Answers (Title+Question, Answer), (Title, Answer), (Title, Question), (Question, Answer) pairs from Yahoo Answers
MS MARCO Triplets (msmarco-distilbert-base-tas-b) (Question, Answer, Negative) triplets from MS MARCO Passages dataset with mined negatives
MS MARCO Triplets (msmarco-distilbert-base-v3) (Question, Answer, Negative) triplets from MS MARCO Passages dataset with mined negatives
MS MARCO Triplets (msmarco-MiniLM-L-6-v3) (Question, Answer, Negative) triplets from MS MARCO Passages dataset with mined negatives
MS MARCO Triplets (distilbert-margin-mse-cls-dot-v2) (Question, Answer, Negative) triplets from MS MARCO Passages dataset with mined negatives
MS MARCO Triplets (distilbert-margin-mse-cls-dot-v1) (Question, Answer, Negative) triplets from MS MARCO Passages dataset with mined negatives
MS MARCO Triplets (distilbert-margin-mse-mean-dot-v1) (Question, Answer, Negative) triplets from MS MARCO Passages dataset with mined negatives
MS MARCO Triplets (mpnet-margin-mse-mean-v1) (Question, Answer, Negative) triplets from MS MARCO Passages dataset with mined negatives
MS MARCO Triplets (co-condenser-margin-mse-cls-v1) (Question, Answer, Negative) triplets from MS MARCO Passages dataset with mined negatives
MS MARCO Triplets (distilbert-margin-mse-mnrl-mean-v1) (Question, Answer, Negative) triplets from MS MARCO Passages dataset with mined negatives
MS MARCO Triplets (distilbert-margin-mse-sym-mnrl-mean-v1) (Question, Answer, Negative) triplets from MS MARCO Passages dataset with mined negatives
MS MARCO Triplets (distilbert-margin-mse-sym-mnrl-mean-v2) (Question, Answer, Negative) triplets from MS MARCO Passages dataset with mined negatives
MS MARCO Triplets (co-condenser-margin-mse-sym-mnrl-mean-v1) (Question, Answer, Negative) triplets from MS MARCO Passages dataset with mined negatives
MS MARCO Triplets (BM25) (Question, Answer, Negative) triplets from MS MARCO Passages dataset with mined negatives
Stack Exchange Duplicates (Title, Title), (Title+Body, Title+Body), (Body, Body) pairs of duplicate questions from StackExchange
ELI5 (Question, Answer) pairs from ELI5 dataset
SQuAD (Question, Answer) pairs from SQuAD dataset
WikiHow (Summary, Text) pairs from WikiHow
Amazon Reviews 2018 (Title, review) pairs from Amazon Reviews
Natural Questions (Query, Answer) pairs from the Natural Questions dataset
Amazon QA (Question, Answer) pairs from Amazon
S2ORC (Title, Abstract), (Abstract, Citation), (Title, Citation) pairs of scientific papers
Quora Duplicates Duplicate question pairs from Quora
WikiAnswers Duplicate question pairs from WikiAnswers
AGNews (Title, Description) pairs of news articles from the AG News dataset
AllNLI (Anchor, Entailment, Contradiction) triplets from SNLI + MultiNLI
NPR (Title, Body) pairs from the npr.org website
SPECTER (Title, Positive Title, Negative Title) triplets of Scientific Publications from Specter
Simple Wiki (English, Simple English) pairs from Wikipedia
PAQ (Query, Answer) from the Probably-Asked Questions dataset
altlex (English, Simple English) pairs from Wikipedia
CC News (Title, article) pairs from the CC News dataset
CodeSearchNet (Comment, Code) pairs from open source libraries on GitHub
Sentence Compression (Long text, Short text) pairs from the Sentence Compression dataset
Trivia QA (Query, Answer) pairs from the TriviaQA dataset
Flickr30k Captions Duplicate captions from the Flickr30k dataset
xsum (News Article, Summary) pairs from XSUM dataset
Coco Captions Duplicate captions from the Coco Captions dataset
Parallel Sentences: Europarl (English, Non-English) pairs across numerous languages
Parallel Sentences: Global Voices (English, Non-English) pairs across numerous languages
Parallel Sentences: MUSE (English, Non-English) pairs across numerous languages
Parallel Sentences: JW300 (English, Non-English) pairs across numerous languages
Parallel Sentences: News Commentary (English, Non-English) pairs across numerous languages
Parallel Sentences: OpenSubtitles (English, Non-English) pairs across numerous languages
Parallel Sentences: Talks (English, Non-English) pairs across numerous languages
Parallel Sentences: Tatoeba (English, Non-English) pairs across numerous languages
Parallel Sentences: WikiMatrix (English, Non-English) pairs across numerous languages
Parallel Sentences: WikiTitles (English, Non-English) pairs across numerous languages

Note

We advise users to tag datasets that can be used for training embedding models with sentence-transformers by adding tags: sentence-transformers. We would also gladly accept high quality datasets to be added to the list above for all to see and use.