{ "results": [ { "task_name": "GEM/web_nlg_en", "prompt_name": "PALM_prompt", "bleu": 0.31606716251794437, "dataset_path": "GEM/web_nlg", "dataset_name": "en", "subset": null, "bleu_stderr": 0.02563944351471059 }, { "task_name": "GEM/web_nlg_en", "prompt_name": "PALM_prompt", "rouge1_precision": 0.06664761750887295, "dataset_path": "GEM/web_nlg", "dataset_name": "en", "subset": null, "rouge1_precision_stderr": 0.0013732959148702218 }, { "task_name": "GEM/web_nlg_en", "prompt_name": "PALM_prompt", "rouge1_recall": 0.2940457093239289, "dataset_path": "GEM/web_nlg", "dataset_name": "en", "subset": null, "rouge1_recall_stderr": 0.004555499438691762 }, { "task_name": "GEM/web_nlg_en", "prompt_name": "PALM_prompt", "rouge1_fmeasure": 0.10246108718545202, "dataset_path": "GEM/web_nlg", "dataset_name": "en", "subset": null, "rouge1_fmeasure_stderr": 0.0017992092481067788 }, { "task_name": "GEM/web_nlg_en", "prompt_name": "PALM_prompt", "rouge2_precision": 0.030425492420252335, "dataset_path": "GEM/web_nlg", "dataset_name": "en", "subset": null, "rouge2_precision_stderr": 0.0007598510657945697 }, { "task_name": "GEM/web_nlg_en", "prompt_name": "PALM_prompt", "rouge2_recall": 0.1409722690208349, "dataset_path": "GEM/web_nlg", "dataset_name": "en", "subset": null, "rouge2_recall_stderr": 0.0031027610869315144 }, { "task_name": "GEM/web_nlg_en", "prompt_name": "PALM_prompt", "rouge2_fmeasure": 0.04746695901782121, "dataset_path": "GEM/web_nlg", "dataset_name": "en", "subset": null, "rouge2_fmeasure_stderr": 0.0011129470648980045 }, { "task_name": "GEM/web_nlg_en", "prompt_name": "PALM_prompt", "rougeL_precision": 0.06400114097684483, "dataset_path": "GEM/web_nlg", "dataset_name": "en", "subset": null, "rougeL_precision_stderr": 0.0012983914696295615 }, { "task_name": "GEM/web_nlg_en", "prompt_name": "PALM_prompt", "rougeL_recall": 0.2849573873935468, "dataset_path": "GEM/web_nlg", "dataset_name": "en", "subset": null, "rougeL_recall_stderr": 0.004452046846530438 }, { "task_name": "GEM/web_nlg_en", "prompt_name": "PALM_prompt", "rougeL_fmeasure": 0.09860897831373452, "dataset_path": "GEM/web_nlg", "dataset_name": "en", "subset": null, "rougeL_fmeasure_stderr": 0.00170202692564 }, { "task_name": "GEM/web_nlg_en", "prompt_name": "PALM_prompt", "rougeLsum_precision": 0.06366499205115346, "dataset_path": "GEM/web_nlg", "dataset_name": "en", "subset": null, "rougeLsum_precision_stderr": 0.0013067510671992333 }, { "task_name": "GEM/web_nlg_en", "prompt_name": "PALM_prompt", "rougeLsum_recall": 0.281020892101209, "dataset_path": "GEM/web_nlg", "dataset_name": "en", "subset": null, "rougeLsum_recall_stderr": 0.004305691903165883 }, { "task_name": "GEM/web_nlg_en", "prompt_name": "PALM_prompt", "rougeLsum_fmeasure": 0.09784615421581211, "dataset_path": "GEM/web_nlg", "dataset_name": "en", "subset": null, "rougeLsum_fmeasure_stderr": 0.001694890791053129 } ], "config": { "model": "hf-causal", "model_args": "pretrained=/pfs/lustrep4/scratch/project_462000119/muennighoff/nov-2022-bettercom/lm1-8b7-178b-c4-repetitions/8b7178b25b/transformers,use_accelerate=True,tokenizer=/pfs/lustrep4/scratch/project_462000119/muennighoff/nov-2022-bettercom/gpt2,dtype=bfloat16", "task_args": "", "num_fewshot": 0, "batch_size": 16, "device": "cuda", "use_cache": false, "limit": 3000, "bootstrap_iters": 10, "seed": 1234 } }