Muennighoff's picture
Add MTEB evaluation
154c4e9
raw
history blame
939 Bytes
{
"dataset_version": null,
"mteb_version": "0.0.2",
"test": {
"evaluation_time": 64.44,
"map_at_1": 0.22139,
"map_at_10": 0.28839,
"map_at_100": 0.30023,
"map_at_1000": 0.30153,
"map_at_3": 0.26521,
"map_at_5": 0.27775,
"mrr_at_1": 0.26466,
"mrr_at_10": 0.33495,
"mrr_at_100": 0.34417,
"mrr_at_1000": 0.34485,
"mrr_at_3": 0.31402,
"mrr_at_5": 0.32496,
"ndcg_at_1": 0.26466,
"ndcg_at_10": 0.33372,
"ndcg_at_100": 0.387,
"ndcg_at_1000": 0.41696,
"ndcg_at_3": 0.29443,
"ndcg_at_5": 0.31121,
"precision_at_1": 0.26466,
"precision_at_10": 0.06037,
"precision_at_100": 0.01067,
"precision_at_1000": 0.00162,
"precision_at_3": 0.13782,
"precision_at_5": 0.09757,
"recall_at_1": 0.22139,
"recall_at_10": 0.4239,
"recall_at_100": 0.65427,
"recall_at_1000": 0.86049,
"recall_at_3": 0.31127,
"recall_at_5": 0.35718
}
}