SOTAVerified

SGPT: GPT Sentence Embeddings for Semantic Search

2022-02-17Code Available2· sign in to hype

Niklas Muennighoff

Code Available — Be the first to reproduce this paper.

Reproduce

Code

Abstract

Decoder transformers have continued increasing in scale reaching hundreds of billions of parameters. Due to their scale the same decoder sets state-of-the-art results on various language tasks via prompting or fine-tuning. Yet, these large foundation models remain unusable for the related fields of semantic search and sentence embeddings. This prevents possibly new state-of-the-art results and forces organizations to train and maintain separate models. To this end, we propose SGPT to use decoders for sentence embeddings and semantic search via prompting or fine-tuning. At 5.8 billion parameters SGPT improves on the previously best sentence embeddings by a margin of 7% and outperforms a concurrent method with 175 billion parameters as measured on the BEIR search benchmark. Code, models and result files are freely available at https://github.com/Muennighoff/sgpt.

Tasks

Benchmark Results

DatasetModelMetricClaimedVerifiedStatus
BioASQ (BEIR)SGPT-BE-5.8BnDCG@100.41Unverified
BioASQ (BEIR)SGPT-CE-6.1BnDCG@100.55Unverified
BioASQ (BEIR)SGPT-CE-2.7BnDCG@100.55Unverified
NFCorpus (BEIR)SGPT-CE-2.7BnDCG@100.33Unverified
NFCorpus (BEIR)SGPT-BE-5.8BnDCG@100.36Unverified
NFCorpus (BEIR)OpenAI Search-DavincinDCG@100.36Unverified
NFCorpus (BEIR)SGPT-CE-6.1BnDCG@100.35Unverified
TREC-COVID (BEIR)SGPT-CE-6.1BnDCG@100.79Unverified
TREC-COVID (BEIR)SGPT-CE-2.7BnDCG@100.76Unverified
TREC-COVID (BEIR)SGPT-BE-5.8BnDCG@100.87Unverified

Reproductions