Muennighoff's picture
Add MTEB evaluation
039f451
raw
history blame
939 Bytes
{
"dataset_version": null,
"mteb_version": "0.0.2",
"test": {
"evaluation_time": 65632.31,
"map_at_1": 0.6274,
"map_at_10": 0.7307,
"map_at_100": 0.73398,
"map_at_1000": 0.7341,
"map_at_3": 0.71338,
"map_at_5": 0.72423,
"mrr_at_1": 0.67777,
"mrr_at_10": 0.77873,
"mrr_at_100": 0.78091,
"mrr_at_1000": 0.78094,
"mrr_at_3": 0.76375,
"mrr_at_5": 0.77316,
"ndcg_at_1": 0.67777,
"ndcg_at_10": 0.7824,
"ndcg_at_100": 0.79557,
"ndcg_at_1000": 0.79814,
"ndcg_at_3": 0.75125,
"ndcg_at_5": 0.76834,
"precision_at_1": 0.67777,
"precision_at_10": 0.09832,
"precision_at_100": 0.01061,
"precision_at_1000": 0.0011,
"precision_at_3": 0.29433,
"precision_at_5": 0.18665,
"recall_at_1": 0.6274,
"recall_at_10": 0.89505,
"recall_at_100": 0.95102,
"recall_at_1000": 0.96825,
"recall_at_3": 0.81028,
"recall_at_5": 0.85281
}
}