Muennighoff's picture
Add
b1799ed
{"GEM/web_nlg_en": {"PALM_prompt": {"bleu": 0.11694864689561944, "bleu_stderr": 0.010760323698374666, "rouge1_fmeasure": 0.07558454105004475, "rouge1_fmeasure_stderr": 0.0017728413797174199, "rouge1_precision": 0.05492023407017857, "rouge1_precision_stderr": 0.0016326725952034475, "rouge1_recall": 0.17990254142984613, "rouge1_recall_stderr": 0.0037492925899754617, "rouge2_fmeasure": 0.02956013283155878, "rouge2_fmeasure_stderr": 0.0009596682875711376, "rouge2_precision": 0.02075424080508391, "rouge2_precision_stderr": 0.0009150263161295265, "rouge2_recall": 0.0760260739680381, "rouge2_recall_stderr": 0.0023774787682776578, "rougeL_fmeasure": 0.07194166497021176, "rougeL_fmeasure_stderr": 0.001647469505361452, "rougeL_precision": 0.05153123991260807, "rougeL_precision_stderr": 0.0014581185184364613, "rougeL_recall": 0.17447931108768522, "rougeL_recall_stderr": 0.0036741060344485967, "rougeLsum_fmeasure": 0.07220041477281516, "rougeLsum_fmeasure_stderr": 0.0016608929228483667, "rougeLsum_precision": 0.052002469160781814, "rougeLsum_precision_stderr": 0.0015071594845800093, "rougeLsum_recall": 0.17445084715749506, "rougeLsum_recall_stderr": 0.0036515775079597516}}, "GEM/wiki_lingua_en": {"tldr_en": {"bleu": 0.7583577300885451, "bleu_stderr": 0.0594734221378448, "rouge1_fmeasure": 0.11194017044626671, "rouge1_fmeasure_stderr": 0.0014428492038973415, "rouge1_precision": 0.11992316681250555, "rouge1_precision_stderr": 0.0017675704654899571, "rouge1_recall": 0.13948810001589107, "rouge1_recall_stderr": 0.002058649220725351, "rouge2_fmeasure": 0.010243977830737588, "rouge2_fmeasure_stderr": 0.000493265280559601, "rouge2_precision": 0.010522981314992496, "rouge2_precision_stderr": 0.0006069518989281641, "rouge2_recall": 0.014385796237686857, "rouge2_recall_stderr": 0.0008041815234389601, "rougeL_fmeasure": 0.0897361308600908, "rougeL_fmeasure_stderr": 0.0010822617010161773, "rougeL_precision": 0.09634769905116539, "rougeL_precision_stderr": 0.0013910007655887457, "rougeL_recall": 0.11364453759174203, "rougeL_recall_stderr": 0.0016847934987669815, "rougeLsum_fmeasure": 0.10607719694851259, "rougeLsum_fmeasure_stderr": 0.0013391525579239643, "rougeLsum_precision": 0.11382504495028803, "rougeLsum_precision_stderr": 0.0016623075783859457, "rougeLsum_recall": 0.1321927492875554, "rougeLsum_recall_stderr": 0.0019214290897240306}}, "e2e_nlg_cleaned": {"generate_text_restaurant": {"bleu": 4.290890219148563, "bleu_stderr": 0.11797671966393032, "rouge1_fmeasure": 0.2978425731750224, "rouge1_fmeasure_stderr": 0.0020372027980421125, "rouge1_precision": 0.3459713718406746, "rouge1_precision_stderr": 0.002718482058174072, "rouge1_recall": 0.28815608021642963, "rouge1_recall_stderr": 0.0022672787761088384, "rouge2_fmeasure": 0.08070496923908829, "rouge2_fmeasure_stderr": 0.001400104340028372, "rouge2_precision": 0.09498590594242383, "rouge2_precision_stderr": 0.0017360545885379483, "rouge2_recall": 0.07791865939566535, "rouge2_recall_stderr": 0.0014093737049867404, "rougeL_fmeasure": 0.21768665558633662, "rougeL_fmeasure_stderr": 0.0015933877730589948, "rougeL_precision": 0.2542525441295202, "rougeL_precision_stderr": 0.0021937081691301894, "rougeL_recall": 0.2106927728428706, "rougeL_recall_stderr": 0.0017758508939911513, "rougeLsum_fmeasure": 0.24573333528025115, "rougeLsum_fmeasure_stderr": 0.0018602140974894146, "rougeLsum_precision": 0.2862920446603683, "rougeLsum_precision_stderr": 0.002475870560064218, "rougeLsum_recall": 0.2376715264121337, "rougeLsum_recall_stderr": 0.002037562231835293}}, "gem_xsum": {"article_DOC_summary": {"bleu": 0.5986931718476021, "bleu_stderr": 0.08317206014621986, "rouge1_fmeasure": 0.12357434836180202, "rouge1_fmeasure_stderr": 0.002149017836427172, "rouge1_precision": 0.08876555984023238, "rouge1_precision_stderr": 0.0016308308833255663, "rouge1_recall": 0.21454525270713967, "rouge1_recall_stderr": 0.0035861819885292467, "rouge2_fmeasure": 0.01649554595160008, "rouge2_fmeasure_stderr": 0.000971551127185768, "rouge2_precision": 0.011772595475451241, "rouge2_precision_stderr": 0.0007152713643140575, "rouge2_recall": 0.02914764125397616, "rouge2_recall_stderr": 0.001651586800843389, "rougeL_fmeasure": 0.10362842140926162, "rougeL_fmeasure_stderr": 0.0016508490182790031, "rougeL_precision": 0.07425164547734261, "rougeL_precision_stderr": 0.0012476815962285165, "rougeL_recall": 0.18105447617675274, "rougeL_recall_stderr": 0.002824838892161731, "rougeLsum_fmeasure": 0.09716065684449275, "rougeLsum_fmeasure_stderr": 0.0017475396647593806, "rougeLsum_precision": 0.06960818379765893, "rougeLsum_precision_stderr": 0.0013153158283565624, "rougeLsum_recall": 0.17002840065420005, "rougeLsum_recall_stderr": 0.002984809197834264}}}