Muennighoff's picture
Add
8393ff0
{"GEM/web_nlg_en": {"PALM_prompt": {"bleu": 0.22225022702141145, "bleu_stderr": 0.02621856502432845, "rouge1_fmeasure": 0.07736875013744622, "rouge1_fmeasure_stderr": 0.001691490572090287, "rouge1_precision": 0.0550474866122264, "rouge1_precision_stderr": 0.0017381938887866828, "rouge1_recall": 0.20167791501798338, "rouge1_recall_stderr": 0.0038251546295238435, "rouge2_fmeasure": 0.032417847332262266, "rouge2_fmeasure_stderr": 0.001002648509701461, "rouge2_precision": 0.021397614706184453, "rouge2_precision_stderr": 0.0007281130169818331, "rouge2_recall": 0.090645566310036, "rouge2_recall_stderr": 0.0026025344432351665, "rougeL_fmeasure": 0.07451634972191427, "rougeL_fmeasure_stderr": 0.0015994885127526045, "rougeL_precision": 0.052662428071746134, "rougeL_precision_stderr": 0.0016500714413103432, "rougeL_recall": 0.19657742260229802, "rougeL_recall_stderr": 0.003721411425073946, "rougeLsum_fmeasure": 0.07461229148174212, "rougeLsum_fmeasure_stderr": 0.0016039862328925923, "rougeLsum_precision": 0.05293413097726497, "rougeLsum_precision_stderr": 0.0016741857992613612, "rougeLsum_recall": 0.19616793121153325, "rougeLsum_recall_stderr": 0.0036738192017033164}}, "GEM/wiki_lingua_en": {"tldr_en": {"bleu": 0.3424809589815014, "bleu_stderr": 0.04930397521232032, "rouge1_fmeasure": 0.09820654856900671, "rouge1_fmeasure_stderr": 0.001187853909005181, "rouge1_precision": 0.10992468154838665, "rouge1_precision_stderr": 0.001537175132948309, "rouge1_recall": 0.11526797449183464, "rouge1_recall_stderr": 0.001627718515491491, "rouge2_fmeasure": 0.0046068536521459615, "rouge2_fmeasure_stderr": 0.00029036627707368774, "rouge2_precision": 0.004999666230578097, "rouge2_precision_stderr": 0.00030886992992011755, "rouge2_recall": 0.005921789250512065, "rouge2_recall_stderr": 0.0004794791832839776, "rougeL_fmeasure": 0.07870934391885585, "rougeL_fmeasure_stderr": 0.0008804322165275051, "rougeL_precision": 0.0886160168917866, "rougeL_precision_stderr": 0.0012010769586605278, "rougeL_recall": 0.09333121266439459, "rougeL_recall_stderr": 0.0012821368035368765, "rougeLsum_fmeasure": 0.0945948924904273, "rougeLsum_fmeasure_stderr": 0.0011308323540178784, "rougeLsum_precision": 0.1058389901935815, "rougeLsum_precision_stderr": 0.001467490624430699, "rougeLsum_recall": 0.1111198364220088, "rougeLsum_recall_stderr": 0.0015509373678568702}}, "e2e_nlg_cleaned": {"generate_text_restaurant": {"bleu": 4.366418005896788, "bleu_stderr": 0.09676432897078642, "rouge1_fmeasure": 0.30174526089566334, "rouge1_fmeasure_stderr": 0.002066630447998624, "rouge1_precision": 0.35920566934792436, "rouge1_precision_stderr": 0.0028557714252695553, "rouge1_recall": 0.28855278953684604, "rouge1_recall_stderr": 0.0023454483555028533, "rouge2_fmeasure": 0.08887408125831098, "rouge2_fmeasure_stderr": 0.0014325940548637634, "rouge2_precision": 0.11019762762876824, "rouge2_precision_stderr": 0.0021416631049273127, "rouge2_recall": 0.08474529585703953, "rouge2_recall_stderr": 0.0014288331957721101, "rougeL_fmeasure": 0.21936946201462224, "rougeL_fmeasure_stderr": 0.0016068916407094758, "rougeL_precision": 0.26370320114619583, "rougeL_precision_stderr": 0.002401220393702974, "rougeL_recall": 0.20942666991545814, "rougeL_recall_stderr": 0.0017933860594759126, "rougeLsum_fmeasure": 0.24795004424369418, "rougeLsum_fmeasure_stderr": 0.0018668078494820878, "rougeLsum_precision": 0.2968494161669508, "rougeLsum_precision_stderr": 0.0026480886568767516, "rougeLsum_recall": 0.2366382950376717, "rougeLsum_recall_stderr": 0.002052819175124416}}, "gem_xsum": {"article_DOC_summary": {"bleu": 0.5711155646101658, "bleu_stderr": 0.07781463462993297, "rouge1_fmeasure": 0.13181267442770545, "rouge1_fmeasure_stderr": 0.0021309245057663934, "rouge1_precision": 0.09799309397257797, "rouge1_precision_stderr": 0.0016694211953608505, "rouge1_recall": 0.22152155486684968, "rouge1_recall_stderr": 0.0037159563326858718, "rouge2_fmeasure": 0.016264206880986808, "rouge2_fmeasure_stderr": 0.0008575910429309486, "rouge2_precision": 0.011550119877479167, "rouge2_precision_stderr": 0.000611649499224846, "rouge2_recall": 0.02917576526370157, "rouge2_recall_stderr": 0.001590971286570559, "rougeL_fmeasure": 0.10703573894361465, "rougeL_fmeasure_stderr": 0.00157513587612925, "rougeL_precision": 0.07948692608742222, "rougeL_precision_stderr": 0.0012394269752421976, "rougeL_recall": 0.1807161877991662, "rougeL_recall_stderr": 0.002821961833500358, "rougeLsum_fmeasure": 0.10216325712741849, "rougeLsum_fmeasure_stderr": 0.0016621446930822356, "rougeLsum_precision": 0.07590517466535618, "rougeLsum_precision_stderr": 0.0012981884730369406, "rougeLsum_recall": 0.17289177898109764, "rougeLsum_recall_stderr": 0.002988721521728284}}}