This similarity model was trained for 2 epochs based on the sentence-transformers/all-mpnet-base-v2. We used the SQuAD 2.0 dataset to compare similarity between questions and sentences containing the answer to the question. We employed the MultipleNegativesRankingLoss as the objective function. To generate negative examples, our strategy involved using BM25 to retrieve similar examples from all sentences in the dataset that did not contain the answer.
As a result, we improved the retrieval of correct sentences (containing the answer) for a question, as measured by Mean Reciprocal Rank (MRR@10) metric. We tested on 7 different SQuAD like datasets.
Model | Drop | BioASQ | Hotpot | News | Textbook | NQ | Trivia |
---|---|---|---|---|---|---|---|
Baseline Dense | 0.37 | 0.12 | 0.28 | 0.29 | 0.17 | 0.29 | 0.47 |
Baseline BM25 | 0.36 | 0.15 | 0.34 | 0.33 | 0.18 | 0.18 | 0.47 |
Azure Hybrid Semantic | 0.35 | 0.18 | 0.29 | ||||
This model | 0.39 | 0.16 | 0.34 | 0.36 | 0.19 | 0.28 | 0.49 |
- Downloads last month
- 25
This model does not have enough activity to be deployed to Inference API (serverless) yet. Increase its social
visibility and check back later, or deploy to Inference Endpoints (dedicated)
instead.