Muennighoff's picture
Add
46fa37e
raw
history blame
4.7 kB
{"GEM/web_nlg_en": {"PALM_prompt": {"bleu": 0.3628804751952509, "bleu_stderr": 0.030697623036573713, "rouge1_fmeasure": 0.11476828922367116, "rouge1_fmeasure_stderr": 0.0026467961791753285, "rouge1_precision": 0.0961267294718708, "rouge1_precision_stderr": 0.0035756778382726984, "rouge1_recall": 0.2651303275965803, "rouge1_recall_stderr": 0.004336290740011959, "rouge2_fmeasure": 0.053377636973435796, "rouge2_fmeasure_stderr": 0.0017003683556607652, "rouge2_precision": 0.04606695542045545, "rouge2_precision_stderr": 0.002435369152747432, "rouge2_recall": 0.12773804414027043, "rouge2_recall_stderr": 0.0030595256393317087, "rougeL_fmeasure": 0.10625511635309616, "rougeL_fmeasure_stderr": 0.0022927878716485555, "rougeL_precision": 0.08840953293951447, "rougeL_precision_stderr": 0.0032719436769350317, "rougeL_recall": 0.25168367399117375, "rougeL_recall_stderr": 0.004042457361339116, "rougeLsum_fmeasure": 0.10755206633558502, "rougeLsum_fmeasure_stderr": 0.002368912902363258, "rougeLsum_precision": 0.08984764804925893, "rougeLsum_precision_stderr": 0.0033451889008913353, "rougeLsum_recall": 0.2531173181775374, "rougeLsum_recall_stderr": 0.004068636618695263}}, "GEM/wiki_lingua_en": {"tldr_en": {"bleu": 1.3044533758133054, "bleu_stderr": 0.06430064424081842, "rouge1_fmeasure": 0.11573778498906329, "rouge1_fmeasure_stderr": 0.0018900659369230771, "rouge1_precision": 0.1501172565902317, "rouge1_precision_stderr": 0.0031593685927941042, "rouge1_recall": 0.1386298198031849, "rouge1_recall_stderr": 0.0025537927988113834, "rouge2_fmeasure": 0.020606436669182757, "rouge2_fmeasure_stderr": 0.0007683955479742162, "rouge2_precision": 0.030016074865409702, "rouge2_precision_stderr": 0.001540633151250424, "rouge2_recall": 0.025565907054811987, "rouge2_recall_stderr": 0.0010672977479962505, "rougeL_fmeasure": 0.09475561950458687, "rougeL_fmeasure_stderr": 0.0014941738837980982, "rougeL_precision": 0.12366016418927188, "rougeL_precision_stderr": 0.0026502505591590366, "rougeL_recall": 0.11497952375628652, "rougeL_recall_stderr": 0.00209855151437069, "rougeLsum_fmeasure": 0.1076559317301304, "rougeLsum_fmeasure_stderr": 0.0017483892549788667, "rougeLsum_precision": 0.14032509931486342, "rougeLsum_precision_stderr": 0.002995754926251608, "rougeLsum_recall": 0.12908967220349415, "rougeLsum_recall_stderr": 0.0023791505225697835}}, "e2e_nlg_cleaned": {"generate_text_restaurant": {"bleu": 9.818363799172253, "bleu_stderr": 0.17226123456239884, "rouge1_fmeasure": 0.4034546893935232, "rouge1_fmeasure_stderr": 0.002274900424644977, "rouge1_precision": 0.48456721060581937, "rouge1_precision_stderr": 0.003156770032478704, "rouge1_recall": 0.38202551287777803, "rouge1_recall_stderr": 0.0027776188479039176, "rouge2_fmeasure": 0.17724850345812884, "rouge2_fmeasure_stderr": 0.00188556182539081, "rouge2_precision": 0.21565285663781844, "rouge2_precision_stderr": 0.002482241476825509, "rouge2_recall": 0.16835492969445903, "rouge2_recall_stderr": 0.002016392425868951, "rougeL_fmeasure": 0.30102287822588836, "rougeL_fmeasure_stderr": 0.0019391871209974672, "rougeL_precision": 0.3636715196709673, "rougeL_precision_stderr": 0.0027615651873153417, "rougeL_recall": 0.2848806242948509, "rougeL_recall_stderr": 0.002283931919678651, "rougeLsum_fmeasure": 0.33530867342554627, "rougeLsum_fmeasure_stderr": 0.002186067051824764, "rougeLsum_precision": 0.4034843209730848, "rougeLsum_precision_stderr": 0.002983415075959704, "rougeLsum_recall": 0.3175119867703305, "rougeLsum_recall_stderr": 0.0025607492206994086}}, "gem_xsum": {"article_DOC_summary": {"bleu": 0.5219766512364415, "bleu_stderr": 0.11233664278264263, "rouge1_fmeasure": 0.13622909776520914, "rouge1_fmeasure_stderr": 0.0024088162106530512, "rouge1_precision": 0.1308676353024568, "rouge1_precision_stderr": 0.0028512595715572946, "rouge1_recall": 0.17286924580165905, "rouge1_recall_stderr": 0.0032482545485716607, "rouge2_fmeasure": 0.01296402126935033, "rouge2_fmeasure_stderr": 0.000839803958464097, "rouge2_precision": 0.012221469982448533, "rouge2_precision_stderr": 0.0008376194509775293, "rouge2_recall": 0.01699862941501089, "rouge2_recall_stderr": 0.001148017122137814, "rougeL_fmeasure": 0.10564141113650093, "rougeL_fmeasure_stderr": 0.0017554003131059596, "rougeL_precision": 0.10179740760145482, "rougeL_precision_stderr": 0.0021956947443623623, "rougeL_recall": 0.1346269556609733, "rougeL_recall_stderr": 0.0024186652980628704, "rougeLsum_fmeasure": 0.1076999813649026, "rougeLsum_fmeasure_stderr": 0.0018096662683577843, "rougeLsum_precision": 0.10337434797074006, "rougeLsum_precision_stderr": 0.0022089632165215025, "rougeLsum_recall": 0.137989344212552, "rougeLsum_recall_stderr": 0.00258889918517997}}}