href_results / temperature=1.0 /mpt-7b-chat.json
alrope's picture
Upload folder using huggingface_hub
efa4eea verified
{
"path": "mosaicml/mpt-7b-chat",
"brainstorm": 0.007,
"open_qa": 0.696,
"closed_qa": 0.062,
"extract": 0.02,
"generation": 0.04,
"rewrite": 0.034,
"summarize": 0.0,
"classify": 0.124,
"reasoning_over_numerical_data": 0.037,
"multi-document_synthesis": 0.013,
"fact_checking_or_attributed_qa": 0.32,
"average": 0.0795,
"brainstorm_rank": 22,
"open_qa_rank": 1,
"closed_qa_rank": 21,
"extract_rank": 21,
"generation_rank": 21,
"rewrite_rank": 21,
"summarize_rank": 25,
"classify_rank": 21,
"reasoning_over_numerical_data_rank": 19,
"multi-document_synthesis_rank": 23,
"fact_checking_or_attributed_qa_rank": 15,
"average_rank": 22,
"brainstorm_confi": "+0.7 / -0.5",
"open_qa_confi": "+8.8 / -9.8",
"closed_qa_confi": "+3.2 / -3.0",
"extract_confi": "+2.0 / -1.5",
"generation_confi": "+1.7 / -1.5",
"rewrite_confi": "+1.4 / -1.3",
"summarize_confi": "+0.0 / -0.0",
"classify_confi": "+5.0 / -4.2",
"reasoning_over_numerical_data_confi": "+1.7 / -1.5",
"multi-document_synthesis_confi": "+1.1 / -0.9",
"fact_checking_or_attributed_qa_confi": "+4.3 / -4.3",
"average_confi": "+0.82 / -0.76"
}