This similarity model was trained for 2 epochs based on the sentence-transformers/all-mpnet-base-v2. We used the SQuAD 2.0 dataset to compare similarity between questions and sentences containing the answer to the question. We employed the MultipleNegativesRankingLoss as the objective function. To generate negative examples, our strategy involved using BM25 to retrieve similar examples from all sentences in the dataset that did not contain the answer.

As a result, we improved the retrieval of correct sentences (containing the answer) for a question, as measured by Mean Reciprocal Rank (MRR@10) metric. We tested on 7 different SQuAD like datasets.

Model Drop BioASQ Hotpot News Textbook NQ Trivia
Baseline Dense 0.37 0.12 0.28 0.29 0.17 0.29 0.47
Baseline BM25 0.36 0.15 0.34 0.33 0.18 0.18 0.47
Azure Hybrid Semantic 0.35 0.18 0.29
This model 0.39 0.16 0.34 0.36 0.19 0.28 0.49
Downloads last month
27
Inference Examples
This model does not have enough activity to be deployed to Inference API (serverless) yet. Increase its social visibility and check back later, or deploy to Inference Endpoints (dedicated) instead.

Dataset used to train vabatista/sbert-mpnet-base-bm25-hard-neg