This model is DPR trained on MS MARCO. The training details and evaluation results are as follows:

Model Pretrain Model Train w/ Marco Title Marco Dev MRR@10 BEIR Avg NDCG@10
DPR bert-base-uncased w/ 32.4 35.5
BERI Dataset NDCG@10
TREC-COVID 58.8
NFCorpus 23.4
FiQA 20.6
ArguAna 39.4
Touché-2020 22.3
Quora 78.0
SCIDOCS 11.9
SciFact 49.4
NQ 43.9
HotpotQA 45.3
Signal-1M 20.2
TREC-NEWS 31.8
DBPedia-entity 28.7
Fever 65.0
Climate-Fever 14.9
BioASQ 24.1
Robust04 32.3
CQADupStack 28.3

The implementation is the same as our EMNLP 2022 paper "Reduce Catastrophic Forgetting of Dense Retrieval Training with Teleportation Negatives". The associated GitHub repository is available at https://github.com/OpenMatch/ANCE-Tele.

@inproceedings{sun2022ancetele,
  title={Reduce Catastrophic Forgetting of Dense Retrieval Training with Teleportation Negatives},
  author={Si, Sun and Chenyan, Xiong and Yue, Yu and Arnold, Overwijk and Zhiyuan, Liu and Jie, Bao},
  booktitle={Proceedings of EMNLP 2022},
  year={2022}
}
Downloads last month
8
Inference Providers NEW
This model is not currently available via any of the supported third-party Inference Providers, and the model is not deployed on the HF Inference API.