Muennighoff's picture
Add MTEB evaluation
039f451
raw
history blame
368 Bytes
{
"test": {
"en": {
"accuracy": 0.9346557227542178,
"accuracy_stderr": 0.0072510589372134955,
"f1": 0.9287345917772146,
"f1_stderr": 0.0082627627462556,
"main_score": 0.9346557227542178
},
"evaluation_time": 475.45
},
"dataset_version": null,
"mteb_version": "0.0.2"
}