YAML tags: null
annotations_creators:
- crowdsourced
- expert-generated
languages:
- en
licenses:
- apache-2.0
multilinguality:
- monolingual
pretty_name: P3
size_categories:
- 100M<n<1B
task_categories:
- other
Dataset Card for P3
Table of Contents
Dataset Description
- Homepage: https://bigscience.huggingface.co/promptsource
- Repository: https://github.com/bigscience-workshop/promptsource/
- Paper: TODO
- Point of Contact: Victor Sanh (victor@huggingface.co)
Dataset Summary
P3 (Pubic Pool of Prompts) is a collection of prompted English datasets covering a diverse set of NLP tasks. A prompt is the combination of an input template and a target template. The templates are functions mapping a data example into natural language for the input and target sequences. For example, in the case of an NLI dataset, the data example would include fields for Premise, Hypothesis, Label. An input template would be If {Premise} is true, is it also true that {Hypothesis}?, whereas a target template can be defined with the label choices Choices[label]. Here Choices is prompt-specific metadata that consists of the options yes, maybe, no corresponding to label being entailment (0), neutral (1) or contradiction (2).
Prompts are collected using Promptsource, an interface to interactively write prompts on datasets, and collect prompt-specific metadata such as evaluation metrics. As of October 13th, there are 2'000 prompts collected for 270+ data(sub)sets. The collection of prompts of P3 is publicly available on Promptsource.
To train T0*, we used a subset of the prompts available in Promptsource (see details here). However, some of the prompts use random.choice
, a method that selects uniformly at random an option in a list of valid possibilities. For reproducibility purposes, we release the collection of prompted examples used to train T0*. The data available here are the materialized version of the prompted datasets used in {Multitask Prompted Training Enables Zero-Shot Task Generalization which represent only a subset of the datasets for which there is at least one prompt on Promptsource.
Supported Tasks and Leaderboards
The tasks represented in P3 cover a diverse set of NLP tasks including multiple-choice QA, sentiment analysis or natural language inference. We detail the full list of datasets in Source Data.
Languages
The data in P3 are in English (BCP-47 en
).
Dataset Structure
Data Instances
An example of "train" looks as follows:
{
'answer_choices': ['safe', 'trolley'],
'inputs': [86, 8, 7142, 666, 6, 405, 8, 3, 834, 1518, 21, 1346, 42, 31682, 58, 37, 3, 929, 9, 3042, 63, 2765, 808, 8, 2045, 6448, 326, 13, 8, 31682, 11, 3, 24052, 135, 16, 8, 1346, 552, 8, 3, 834, 47, 6364, 5], 'inputs_pretokenized': 'In the sentence below, does the _ stand for safe or trolley?\nThe treasury workers took the gold bars off of the trolley and stacked them in the safe until the _ was empty.',
'targets': [31682, 1],
'targets_pretokenized': '\ntrolley'
}
In the case of rank classification (letting the model select its the prediction the option with the highest log-likelihood), an example looks as follows:
{
'idx': [5, 0],
'inputs': [86, 8, 7142, 666, 6, 405, 8, 3, 834, 1518, 21, 19454, 42, 22227, 58, 19454, 744, 31, 17, 2112, 4553, 17742, 7, 12, 1953, 6, 298, 22227, 966, 373, 405, 5, 3, 834, 19, 72, 952, 12, 619, 16, 3, 9, 17742, 3298, 5],
'inputs_pretokenized': "In the sentence below, does the _ stand for Kyle or Logan?\nKyle doesn't wear leg warmers to bed, while Logan almost always does. _ is more likely to live in a warmer climate.",
'is_correct': True,
'targets': [19454, 1],
'targets_pretokenized': 'Kyle',
'weight': 1.0
}
To check all the prompted examples, you can use the Promptsource hosted tool and choose the Prompted dataset viewer
mode in the left panel.
Data Fields
The data fields are the same among all splits:
answer_choices
: the choices (in natural language) available to the modelinputs_pretokenized
: the natural language input fed to the modeltargets_pretokenized
: the natural language target that the model has to generateinputs
: the tokenized input with T5's tokenizertargets
: the tokenized target with T5's tokenizeridx
: identifier of the (example, answer_option_id) in the case of rank classificationweight
: a weight for the example produced by seqio (always set to 1.0 in practise)is_correct
: whether the (example, answer_option_id) is the correct one
Data Splits
Data(sub)set | Number of examples per splits |
---|---|
adversarial_qa_dbert_answer_the_following_q | {'train': 10000, 'validation': 1000} |
adversarial_qa_dbert_based_on | {'train': 10000, 'validation': 1000} |
adversarial_qa_dbert_generate_question | {'train': 10000, 'validation': 1000, 'test': 1000} |
adversarial_qa_dbert_question_context_answer | {'train': 10000, 'validation': 1000} |
adversarial_qa_dbert_tell_what_it_is | {'train': 10000, 'validation': 1000} |
adversarial_qa_dbidaf_answer_the_following_q | {'train': 10000, 'validation': 1000} |
adversarial_qa_dbidaf_based_on | {'train': 10000, 'validation': 1000} |
adversarial_qa_dbidaf_generate_question | {'train': 10000, 'validation': 1000, 'test': 1000} |
adversarial_qa_dbidaf_question_context_answer | {'train': 10000, 'validation': 1000} |
adversarial_qa_dbidaf_tell_what_it_is | {'train': 10000, 'validation': 1000} |
adversarial_qa_droberta_answer_the_following_q | {'train': 10000, 'validation': 1000} |
adversarial_qa_droberta_based_on | {'train': 10000, 'validation': 1000} |
adversarial_qa_droberta_generate_question | {'train': 10000, 'validation': 1000, 'test': 1000} |
adversarial_qa_droberta_question_context_answer | {'train': 10000, 'validation': 1000} |
adversarial_qa_droberta_tell_what_it_is | {'train': 10000, 'validation': 1000} |
ag_news_classify | {'train': 120000, 'test': 7600} |
ag_news_classify_question_first | {'train': 120000, 'test': 7600} |
ag_news_classify_with_choices | {'train': 120000, 'test': 7600} |
ag_news_classify_with_choices_question_first | {'train': 120000, 'test': 7600} |
ag_news_recommend | {'train': 120000, 'test': 7600} |
ag_news_which_section | {'train': 120000, 'test': 7600} |
ag_news_which_section_choices | {'train': 120000, 'test': 7600} |
ai2_arc_ARC_Challenge_heres_a_problem | {'train': 1119, 'validation': 299, 'test': 1172} |
ai2_arc_ARC_Challenge_i_am_hesitating | {'train': 1119, 'validation': 299, 'test': 1172} |
ai2_arc_ARC_Challenge_multiple_choice | {'train': 1119, 'validation': 299, 'test': 1172} |
ai2_arc_ARC_Challenge_pick_false_options | {'train': 1119, 'validation': 299, 'test': 1172} |
ai2_arc_ARC_Challenge_pick_the_most_correct_option | {'train': 1119, 'validation': 299, 'test': 1172} |
ai2_arc_ARC_Challenge_qa_options | {'train': 1119, 'validation': 299, 'test': 1172} |
ai2_arc_ARC_Easy_heres_a_problem | {'train': 2251, 'validation': 570, 'test': 2376} |
ai2_arc_ARC_Easy_i_am_hesitating | {'train': 2251, 'validation': 570, 'test': 2376} |
ai2_arc_ARC_Easy_multiple_choice | {'train': 2251, 'validation': 570, 'test': 2376} |
ai2_arc_ARC_Easy_pick_false_options | {'train': 2251, 'validation': 570, 'test': 2376} |
ai2_arc_ARC_Easy_pick_the_most_correct_option | {'train': 2251, 'validation': 570, 'test': 2376} |
ai2_arc_ARC_Easy_qa_options | {'train': 2251, 'validation': 570, 'test': 2376} |
amazon_polarity_Is_this_product_review_positive | {'train': 3600000, 'test': 400000} |
amazon_polarity_Is_this_review | {'train': 3600000, 'test': 400000} |
amazon_polarity_Is_this_review_negative | {'train': 3600000, 'test': 400000} |
amazon_polarity_User_recommend_this_product | {'train': 3600000, 'test': 400000} |
amazon_polarity_convey_negative_or_positive_sentiment | {'train': 3600000, 'test': 400000} |
amazon_polarity_flattering_or_not | {'train': 3600000, 'test': 400000} |
amazon_polarity_negative_or_positive_tone | {'train': 3600000, 'test': 400000} |
amazon_polarity_user_satisfied | {'train': 3600000, 'test': 400000} |
amazon_polarity_would_you_buy | {'train': 3600000, 'test': 400000} |
anli_GPT_3_style_r1 | {'train': 16946, 'validation': 1000, 'test': 1000} |
anli_GPT_3_style_r1_score_eval | {'train': 50838, 'validation': 3000, 'test': 3000} |
anli_GPT_3_style_r2 | {'train': 45460, 'validation': 1000, 'test': 1000} |
anli_GPT_3_style_r2_score_eval | {'train': 136380, 'validation': 3000, 'test': 3000} |
anli_GPT_3_style_r3 | {'train': 100459, 'validation': 1200, 'test': 1200} |
anli_GPT_3_style_r3_score_eval | {'train': 301377, 'validation': 3600, 'test': 3600} |
anli_MNLI_crowdsource_r1 | {'train': 16946, 'validation': 1000, 'test': 1000} |
anli_MNLI_crowdsource_r1_score_eval | {'train': 50838, 'validation': 3000, 'test': 3000} |
anli_MNLI_crowdsource_r2 | {'train': 45460, 'validation': 1000, 'test': 1000} |
anli_MNLI_crowdsource_r2_score_eval | {'train': 136380, 'validation': 3000, 'test': 3000} |
anli_MNLI_crowdsource_r3 | {'train': 100459, 'validation': 1200, 'test': 1200} |
anli_MNLI_crowdsource_r3_score_eval | {'train': 301377, 'validation': 3600, 'test': 3600} |
anli_always_sometimes_never_r1 | {'train': 16946, 'validation': 1000, 'test': 1000} |
anli_always_sometimes_never_r1_score_eval | {'train': 50838, 'validation': 3000, 'test': 3000} |
anli_always_sometimes_never_r2 | {'train': 45460, 'validation': 1000, 'test': 1000} |
anli_always_sometimes_never_r2_score_eval | {'train': 136380, 'validation': 3000, 'test': 3000} |
anli_always_sometimes_never_r3 | {'train': 100459, 'validation': 1200, 'test': 1200} |
anli_always_sometimes_never_r3_score_eval | {'train': 301377, 'validation': 3600, 'test': 3600} |
anli_based_on_the_previous_passage_r1 | {'train': 16946, 'validation': 1000, 'test': 1000} |
anli_based_on_the_previous_passage_r1_score_eval | {'train': 50838, 'validation': 3000, 'test': 3000} |
anli_based_on_the_previous_passage_r2 | {'train': 45460, 'validation': 1000, 'test': 1000} |
anli_based_on_the_previous_passage_r2_score_eval | {'train': 136380, 'validation': 3000, 'test': 3000} |
anli_based_on_the_previous_passage_r3 | {'train': 100459, 'validation': 1200, 'test': 1200} |
anli_based_on_the_previous_passage_r3_score_eval | {'train': 301377, 'validation': 3600, 'test': 3600} |
anli_can_we_infer_r1 | {'train': 16946, 'validation': 1000, 'test': 1000} |
anli_can_we_infer_r1_score_eval | {'train': 50838, 'validation': 3000, 'test': 3000} |
anli_can_we_infer_r2 | {'train': 45460, 'validation': 1000, 'test': 1000} |
anli_can_we_infer_r2_score_eval | {'train': 136380, 'validation': 3000, 'test': 3000} |
anli_can_we_infer_r3 | {'train': 100459, 'validation': 1200, 'test': 1200} |
anli_can_we_infer_r3_score_eval | {'train': 301377, 'validation': 3600, 'test': 3600} |
anli_claim_true_false_inconclusive_r1 | {'train': 16946, 'validation': 1000, 'test': 1000} |
anli_claim_true_false_inconclusive_r1_score_eval | {'train': 50838, 'validation': 3000, 'test': 3000} |
anli_claim_true_false_inconclusive_r2 | {'train': 45460, 'validation': 1000, 'test': 1000} |
anli_claim_true_false_inconclusive_r2_score_eval | {'train': 136380, 'validation': 3000, 'test': 3000} |
anli_claim_true_false_inconclusive_r3 | {'train': 100459, 'validation': 1200, 'test': 1200} |
anli_claim_true_false_inconclusive_r3_score_eval | {'train': 301377, 'validation': 3600, 'test': 3600} |
anli_consider_always_sometimes_never_r1 | {'train': 16946, 'validation': 1000, 'test': 1000} |
anli_consider_always_sometimes_never_r1_score_eval | {'train': 50838, 'validation': 3000, 'test': 3000} |
anli_consider_always_sometimes_never_r2 | {'train': 45460, 'validation': 1000, 'test': 1000} |
anli_consider_always_sometimes_never_r2_score_eval | {'train': 136380, 'validation': 3000, 'test': 3000} |
anli_consider_always_sometimes_never_r3 | {'train': 100459, 'validation': 1200, 'test': 1200} |
anli_consider_always_sometimes_never_r3_score_eval | {'train': 301377, 'validation': 3600, 'test': 3600} |
anli_does_it_follow_that_r1 | {'train': 16946, 'validation': 1000, 'test': 1000} |
anli_does_it_follow_that_r1_score_eval | {'train': 50838, 'validation': 3000, 'test': 3000} |
anli_does_it_follow_that_r2 | {'train': 45460, 'validation': 1000, 'test': 1000} |
anli_does_it_follow_that_r2_score_eval | {'train': 136380, 'validation': 3000, 'test': 3000} |
anli_does_it_follow_that_r3 | {'train': 100459, 'validation': 1200, 'test': 1200} |
anli_does_it_follow_that_r3_score_eval | {'train': 301377, 'validation': 3600, 'test': 3600} |
anli_does_this_imply_r1 | {'train': 16946, 'validation': 1000, 'test': 1000} |
anli_does_this_imply_r1_score_eval | {'train': 50838, 'validation': 3000, 'test': 3000} |
anli_does_this_imply_r2 | {'train': 45460, 'validation': 1000, 'test': 1000} |
anli_does_this_imply_r2_score_eval | {'train': 136380, 'validation': 3000, 'test': 3000} |
anli_does_this_imply_r3 | {'train': 100459, 'validation': 1200, 'test': 1200} |
anli_does_this_imply_r3_score_eval | {'train': 301377, 'validation': 3600, 'test': 3600} |
anli_guaranteed_possible_impossible_r1 | {'train': 16946, 'validation': 1000, 'test': 1000} |
anli_guaranteed_possible_impossible_r1_score_eval | {'train': 50838, 'validation': 3000, 'test': 3000} |
anli_guaranteed_possible_impossible_r2 | {'train': 45460, 'validation': 1000, 'test': 1000} |
anli_guaranteed_possible_impossible_r2_score_eval | {'train': 136380, 'validation': 3000, 'test': 3000} |
anli_guaranteed_possible_impossible_r3 | {'train': 100459, 'validation': 1200, 'test': 1200} |
anli_guaranteed_possible_impossible_r3_score_eval | {'train': 301377, 'validation': 3600, 'test': 3600} |
anli_guaranteed_true_r1 | {'train': 16946, 'validation': 1000, 'test': 1000} |
anli_guaranteed_true_r1_score_eval | {'train': 50838, 'validation': 3000, 'test': 3000} |
anli_guaranteed_true_r2 | {'train': 45460, 'validation': 1000, 'test': 1000} |
anli_guaranteed_true_r2_score_eval | {'train': 136380, 'validation': 3000, 'test': 3000} |
anli_guaranteed_true_r3 | {'train': 100459, 'validation': 1200, 'test': 1200} |
anli_guaranteed_true_r3_score_eval | {'train': 301377, 'validation': 3600, 'test': 3600} |
anli_justified_in_saying_r1 | {'train': 16946, 'validation': 1000, 'test': 1000} |
anli_justified_in_saying_r1_score_eval | {'train': 50838, 'validation': 3000, 'test': 3000} |
anli_justified_in_saying_r2 | {'train': 45460, 'validation': 1000, 'test': 1000} |
anli_justified_in_saying_r2_score_eval | {'train': 136380, 'validation': 3000, 'test': 3000} |
anli_justified_in_saying_r3 | {'train': 100459, 'validation': 1200, 'test': 1200} |
anli_justified_in_saying_r3_score_eval | {'train': 301377, 'validation': 3600, 'test': 3600} |
anli_must_be_true_r1 | {'train': 16946, 'validation': 1000, 'test': 1000} |
anli_must_be_true_r1_score_eval | {'train': 50838, 'validation': 3000, 'test': 3000} |
anli_must_be_true_r2 | {'train': 45460, 'validation': 1000, 'test': 1000} |
anli_must_be_true_r2_score_eval | {'train': 136380, 'validation': 3000, 'test': 3000} |
anli_must_be_true_r3 | {'train': 100459, 'validation': 1200, 'test': 1200} |
anli_must_be_true_r3_score_eval | {'train': 301377, 'validation': 3600, 'test': 3600} |
anli_should_assume_r1 | {'train': 16946, 'validation': 1000, 'test': 1000} |
anli_should_assume_r1_score_eval | {'train': 50838, 'validation': 3000, 'test': 3000} |
anli_should_assume_r2 | {'train': 45460, 'validation': 1000, 'test': 1000} |
anli_should_assume_r2_score_eval | {'train': 136380, 'validation': 3000, 'test': 3000} |
anli_should_assume_r3 | {'train': 100459, 'validation': 1200, 'test': 1200} |
anli_should_assume_r3_score_eval | {'train': 301377, 'validation': 3600, 'test': 3600} |
anli_take_the_following_as_truth_r1 | {'train': 16946, 'validation': 1000, 'test': 1000} |
anli_take_the_following_as_truth_r1_score_eval | {'train': 50838, 'validation': 3000, 'test': 3000} |
anli_take_the_following_as_truth_r2 | {'train': 45460, 'validation': 1000, 'test': 1000} |
anli_take_the_following_as_truth_r2_score_eval | {'train': 136380, 'validation': 3000, 'test': 3000} |
anli_take_the_following_as_truth_r3 | {'train': 100459, 'validation': 1200, 'test': 1200} |
anli_take_the_following_as_truth_r3_score_eval | {'train': 301377, 'validation': 3600, 'test': 3600} |
app_reviews_categorize_rating_using_review | {'train': 288065} |
app_reviews_convert_to_rating | {'train': 288065} |
app_reviews_convert_to_star_rating | {'train': 288065} |
app_reviews_generate_review | {'train': 288065} |
cnn_dailymail_3.0.0_2_or_3_sentences | {'train': 287113, 'validation': 13368, 'test': 11490} |
cnn_dailymail_3.0.0_generate_story | {'train': 287113, 'validation': 13368, 'test': 11490} |
cnn_dailymail_3.0.0_news_card_view | {'train': 287113, 'validation': 13368, 'test': 11490} |
cnn_dailymail_3.0.0_news_stock | {'train': 287113, 'validation': 13368, 'test': 11490} |
cnn_dailymail_3.0.0_news_summary | {'train': 287113, 'validation': 13368, 'test': 11490} |
cnn_dailymail_3.0.0_spice_up_story | {'train': 287113, 'validation': 13368, 'test': 11490} |
cnn_dailymail_3.0.0_sum_in_brief | {'train': 287113, 'validation': 13368, 'test': 11490} |
cnn_dailymail_3.0.0_tldr_summary | {'train': 287113, 'validation': 13368, 'test': 11490} |
cnn_dailymail_3.0.0_write_an_outline | {'train': 287113, 'validation': 13368, 'test': 11490} |
common_gen_Example_prompt | {'train': 67389, 'validation': 4018, 'test': 1497} |
common_gen_Given_concepts_type_1 | {'train': 67389, 'validation': 4018, 'test': 1497} |
common_gen_Given_concepts_type_2 | {'train': 67389, 'validation': 4018, 'test': 1497} |
common_gen_Put_together | {'train': 67389, 'validation': 4018, 'test': 1497} |
common_gen_choice_in_concept_centric_sentence_generation | {'train': 67389, 'validation': 4018, 'test': 1497} |
common_gen_random_task_template_prompt | {'train': 67389, 'validation': 4018, 'test': 1497} |
common_gen_sentence_to_concepts | {'train': 67389, 'validation': 4018, 'test': 1497} |
common_gen_topic_to_sentence | {'train': 67389, 'validation': 4018, 'test': 1497} |
common_gen_topics_from_the_sentence | {'train': 67389, 'validation': 4018, 'test': 1497} |
cos_e_v1.11_aligned_with_common_sense | {'train': 9741, 'validation': 1221} |
cos_e_v1.11_description_question_option_id | {'train': 9741, 'validation': 1221} |
cos_e_v1.11_description_question_option_text | {'train': 9741, 'validation': 1221} |
cos_e_v1.11_explain_why_human | {'train': 9741, 'validation': 1221} |
cos_e_v1.11_generate_explanation_given_text | {'train': 9741, 'validation': 1221} |
cos_e_v1.11_i_think | {'train': 9741, 'validation': 1221} |
cos_e_v1.11_question_description_option_id | {'train': 9741, 'validation': 1221} |
cos_e_v1.11_question_description_option_text | {'train': 9741, 'validation': 1221} |
cos_e_v1.11_question_option_description_id | {'train': 9741, 'validation': 1221} |
cos_e_v1.11_question_option_description_text | {'train': 9741, 'validation': 1221} |
cos_e_v1.11_rationale | {'train': 9741, 'validation': 1221} |
cosmos_qa_context_answer_to_question | {'train': 25262, 'validation': 2985, 'test': 6963} |
cosmos_qa_context_description_question_answer_id | {'train': 25262, 'validation': 2985, 'test': 6963} |
cosmos_qa_context_description_question_answer_text | {'train': 25262, 'validation': 2985, 'test': 6963} |
cosmos_qa_context_description_question_text | {'train': 25262, 'validation': 2985, 'test': 6963} |
cosmos_qa_context_question_description_answer_id | {'train': 25262, 'validation': 2985, 'test': 6963} |
cosmos_qa_context_question_description_answer_text | {'train': 25262, 'validation': 2985, 'test': 6963} |
cosmos_qa_context_question_description_text | {'train': 25262, 'validation': 2985, 'test': 6963} |
cosmos_qa_description_context_question_answer_id | {'train': 25262, 'validation': 2985, 'test': 6963} |
cosmos_qa_description_context_question_answer_text | {'train': 25262, 'validation': 2985, 'test': 6963} |
cosmos_qa_description_context_question_text | {'train': 25262, 'validation': 2985, 'test': 6963} |
cosmos_qa_no_prompt_id | {'train': 25262, 'validation': 2985, 'test': 6963} |
cosmos_qa_no_prompt_text | {'train': 25262, 'validation': 2985, 'test': 6963} |
cosmos_qa_only_question_answer | {'train': 25262, 'validation': 2985, 'test': 6963} |
dbpedia_14_given_a_choice_of_categories_ | {'train': 560000, 'test': 70000} |
dbpedia_14_given_a_list_of_category_what_does_the_title_belong_to | {'train': 560000, 'test': 70000} |
dbpedia_14_given_list_what_category_does_the_paragraph_belong_to | {'train': 560000, 'test': 70000} |
dbpedia_14_pick_one_category_for_the_following_text | {'train': 560000, 'test': 70000} |
dream_answer_to_dialogue | {'train': 6116, 'validation': 2040, 'test': 2041} |
dream_baseline | {'train': 6116, 'validation': 2040, 'test': 2041} |
dream_generate_first_utterance | {'train': 6116, 'validation': 2040, 'test': 2041} |
dream_generate_last_utterance | {'train': 6116, 'validation': 2040, 'test': 2041} |
dream_read_the_following_conversation_and_answer_the_question | {'train': 6116, 'validation': 2040, 'test': 2041} |
duorc_ParaphraseRC_answer_question | {'train': 69524, 'validation': 15591, 'test': 15857} |
duorc_ParaphraseRC_build_story_around_qa | {'train': 58752, 'validation': 13111, 'test': 13449} |
duorc_ParaphraseRC_decide_worth_it | {'train': 69524, 'validation': 15591, 'test': 15857} |
duorc_ParaphraseRC_extract_answer | {'train': 69524, 'validation': 15591, 'test': 15857} |
duorc_ParaphraseRC_generate_question | {'train': 69524, 'validation': 15591, 'test': 15857} |
duorc_ParaphraseRC_generate_question_by_answer | {'train': 58752, 'validation': 13111, 'test': 13449} |
duorc_ParaphraseRC_movie_director | {'train': 69524, 'validation': 15591, 'test': 15857} |
duorc_ParaphraseRC_question_answering | {'train': 69524, 'validation': 15591, 'test': 15857} |
duorc_ParaphraseRC_title_generation | {'train': 69524, 'validation': 15591, 'test': 15857} |
duorc_SelfRC_answer_question | {'train': 60721, 'validation': 12961, 'test': 12559} |
duorc_SelfRC_build_story_around_qa | {'train': 60094, 'validation': 12845, 'test': 12415} |
duorc_SelfRC_decide_worth_it | {'train': 60721, 'validation': 12961, 'test': 12559} |
duorc_SelfRC_extract_answer | {'train': 60721, 'validation': 12961, 'test': 12559} |
duorc_SelfRC_generate_question | {'train': 60721, 'validation': 12961, 'test': 12559} |
duorc_SelfRC_generate_question_by_answer | {'train': 60094, 'validation': 12845, 'test': 12415} |
duorc_SelfRC_movie_director | {'train': 60721, 'validation': 12961, 'test': 12559} |
duorc_SelfRC_question_answering | {'train': 60721, 'validation': 12961, 'test': 12559} |
duorc_SelfRC_title_generation | {'train': 60721, 'validation': 12961, 'test': 12559} |
gigaword_TLDR | {'train': 3803957, 'validation': 189651, 'test': 1951} |
gigaword_first_sentence_title | {'train': 3803957, 'validation': 189651, 'test': 1951} |
gigaword_generate_summary_for_this | {'train': 3803957, 'validation': 189651, 'test': 1951} |
gigaword_in_a_nutshell | {'train': 3803957, 'validation': 189651, 'test': 1951} |
gigaword_make_a_title | {'train': 3803957, 'validation': 189651, 'test': 1951} |
gigaword_reverse_writing | {'train': 3803957, 'validation': 189651, 'test': 1951} |
gigaword_write_a_title_for_this_sentence | {'train': 3803957, 'validation': 189651, 'test': 1951} |
gigaword_write_an_article | {'train': 3803957, 'validation': 189651, 'test': 1951} |
gigaword_write_its_sentence | {'train': 3803957, 'validation': 189651, 'test': 1951} |
glue_mrpc_equivalent | {'train': 3668, 'validation': 408, 'test': 1725} |
glue_mrpc_generate_paraphrase | {'train': 2474, 'validation': 279, 'test': 1147} |
glue_mrpc_generate_sentence | {'train': 2474, 'validation': 279, 'test': 1147} |
glue_mrpc_paraphrase | {'train': 3668, 'validation': 408, 'test': 1725} |
glue_mrpc_replace | {'train': 3668, 'validation': 408, 'test': 1725} |
glue_mrpc_same_thing | {'train': 3668, 'validation': 408, 'test': 1725} |
glue_mrpc_want_to_know | {'train': 3668, 'validation': 408, 'test': 1725} |
glue_qqp_answer | {'train': 363846, 'validation': 40430, 'test': 390965} |
glue_qqp_duplicate | {'train': 363846, 'validation': 40430, 'test': 390965} |
glue_qqp_duplicate_or_not | {'train': 363846, 'validation': 40430, 'test': 390965} |
glue_qqp_meaning | {'train': 363846, 'validation': 40430, 'test': 390965} |
glue_qqp_quora | {'train': 363846, 'validation': 40430, 'test': 390965} |
glue_qqp_same_thing | {'train': 363846, 'validation': 40430, 'test': 390965} |
hellaswag_Appropriate_continuation_Yes_or_No | {'train': 39905, 'validation': 10042, 'test': 10003} |
hellaswag_Open_ended_completion | {'train': 39905, 'validation': 10042, 'test': 10003} |
hellaswag_Open_ended_start | {'train': 39905, 'validation': 10042, 'test': 10003} |
hellaswag_Predict_ending_with_hint | {'train': 39905, 'validation': 10042, 'test': 10003} |
hellaswag_Predict_ending_with_hint_score_eval | {'train': 159620, 'validation': 40168, 'test': 40012} |
hellaswag_Randomized_prompts_template | {'train': 39905, 'validation': 10042, 'test': 10003} |
hellaswag_Randomized_prompts_template_score_eval | {'train': 159620, 'validation': 40168, 'test': 40012} |
hellaswag_Reversed_appropriate_continuation_Yes_or_No | {'train': 39905, 'validation': 10042, 'test': 10003} |
hellaswag_Topic_of_the_context | {'train': 39905, 'validation': 10042, 'test': 10003} |
hellaswag_Topic_without_the_ending_answer | {'train': 39905, 'validation': 10042, 'test': 10003} |
hellaswag_complete_first_then | {'train': 39905, 'validation': 10042, 'test': 10003} |
hellaswag_complete_first_then_score_eval | {'train': 159620, 'validation': 40168, 'test': 40012} |
hellaswag_how_ends | {'train': 39905, 'validation': 10042, 'test': 10003} |
hellaswag_if_begins_how_continues | {'train': 39905, 'validation': 10042, 'test': 10003} |
hellaswag_if_begins_how_continues_score_eval | {'train': 159620, 'validation': 40168, 'test': 40012} |
imdb_Movie_Expressed_Sentiment | {'train': 25000, 'test': 25000, 'unsupervised': 50000} |
imdb_Movie_Expressed_Sentiment_2 | {'train': 25000, 'test': 25000, 'unsupervised': 50000} |
imdb_Negation_template_for_positive_and_negative | {'train': 25000, 'test': 25000, 'unsupervised': 50000} |
imdb_Reviewer_Enjoyment | {'train': 25000, 'test': 25000, 'unsupervised': 50000} |
imdb_Reviewer_Enjoyment_Yes_No | {'train': 25000, 'test': 25000, 'unsupervised': 50000} |
imdb_Reviewer_Expressed_Sentiment | {'train': 25000, 'test': 25000, 'unsupervised': 50000} |
imdb_Reviewer_Opinion_bad_good_choices | {'train': 25000, 'test': 25000, 'unsupervised': 50000} |
imdb_Reviewer_Sentiment_Feeling | {'train': 25000, 'test': 25000, 'unsupervised': 50000} |
imdb_Sentiment_with_choices_ | {'train': 25000, 'test': 25000, 'unsupervised': 50000} |
imdb_Text_Expressed_Sentiment | {'train': 25000, 'test': 25000, 'unsupervised': 50000} |
imdb_Writer_Expressed_Sentiment | {'train': 25000, 'test': 25000, 'unsupervised': 50000} |
kilt_tasks_hotpotqa_combining_facts | {'train': 88869, 'validation': 5600} |
kilt_tasks_hotpotqa_complex_question | {'train': 88869, 'validation': 5600} |
kilt_tasks_hotpotqa_final_exam | {'train': 88869, 'validation': 5600} |
kilt_tasks_hotpotqa_formulate | {'train': 88869, 'validation': 5600} |
kilt_tasks_hotpotqa_straighforward_qa | {'train': 88869, 'validation': 5600} |
multi_news_distill | {'train': 44972, 'validation': 5622, 'test': 5622} |
multi_news_expand_reverse_task_ | {'train': 44972, 'validation': 5622, 'test': 5622} |
multi_news_summarize | {'train': 44972, 'validation': 5622, 'test': 5622} |
multi_news_summary_scenario | {'train': 44972, 'validation': 5622, 'test': 5622} |
multi_news_synthesize | {'train': 44972, 'validation': 5622, 'test': 5622} |
multi_news_what_are_the_key_points | {'train': 44972, 'validation': 5622, 'test': 5622} |
openbookqa_main_choices | {'train': 4957, 'validation': 500, 'test': 500} |
openbookqa_main_choose_an_answer_with_options | {'train': 4957, 'validation': 500, 'test': 500} |
openbookqa_main_only_options | {'train': 4957, 'validation': 500, 'test': 500} |
openbookqa_main_pick_answer_with_options | {'train': 4957, 'validation': 500, 'test': 500} |
openbookqa_main_pick_using_id | {'train': 4957, 'validation': 500, 'test': 500} |
openbookqa_main_which_correct | {'train': 4957, 'validation': 500, 'test': 500} |
openbookqa_main_which_correct_inverse | {'train': 4957, 'validation': 500, 'test': 500} |
paws_labeled_final_Concatenation | {'train': 49401, 'validation': 8000, 'test': 8000} |
paws_labeled_final_Concatenation_no_label | {'train': 49401, 'validation': 8000, 'test': 8000} |
paws_labeled_final_Meaning | {'train': 49401, 'validation': 8000, 'test': 8000} |
paws_labeled_final_Meaning_no_label | {'train': 49401, 'validation': 8000, 'test': 8000} |
paws_labeled_final_PAWS_ANLI_GPT3 | {'train': 49401, 'validation': 8000, 'test': 8000} |
paws_labeled_final_PAWS_ANLI_GPT3_no_label | {'train': 49401, 'validation': 8000, 'test': 8000} |
paws_labeled_final_Rewrite | {'train': 49401, 'validation': 8000, 'test': 8000} |
paws_labeled_final_Rewrite_no_label | {'train': 49401, 'validation': 8000, 'test': 8000} |
paws_labeled_final_context_question | {'train': 49401, 'validation': 8000, 'test': 8000} |
paws_labeled_final_context_question_no_label | {'train': 49401, 'validation': 8000, 'test': 8000} |
paws_labeled_final_paraphrase_task | {'train': 21829, 'validation': 3539, 'test': 3536} |
paws_labeled_final_task_description_no_label | {'train': 49401, 'validation': 8000, 'test': 8000} |
piqa_Correct_the_solution | {'train': 16113, 'validation': 1838, 'test': 3084} |
piqa_Correct_the_solution_if_false_from_sol_1 | {'train': 16113, 'validation': 1838, 'test': 3084} |
piqa_Correct_the_solution_if_false_from_sol_2 | {'train': 16113, 'validation': 1838, 'test': 3084} |
piqa_Does_this_solution_make_sense_sol1 | {'train': 16113, 'validation': 1838, 'test': 3084} |
piqa_Does_this_solution_make_sense_sol2 | {'train': 16113, 'validation': 1838, 'test': 3084} |
piqa_choose_the_most_appropriate_solution | {'train': 16113, 'validation': 1838, 'test': 3084} |
piqa_finish_sentence_with_correct_choice | {'train': 16113, 'validation': 1838, 'test': 3084} |
piqa_no_prompt_needed | {'train': 16113, 'validation': 1838, 'test': 3084} |
piqa_pick_correct_choice_index | {'train': 16113, 'validation': 1838, 'test': 3084} |
piqa_pick_correct_choice_with_choice_given_before_goal | {'train': 16113, 'validation': 1838, 'test': 3084} |
piqa_what_is_the_correct_ending | {'train': 16113, 'validation': 1838, 'test': 3084} |
qasc_is_correct_1 | {'train': 8134, 'validation': 926, 'test': 920} |
qasc_is_correct_2 | {'train': 8134, 'validation': 926, 'test': 920} |
qasc_qa_with_combined_facts_1 | {'train': 8134, 'validation': 926, 'test': 920} |
qasc_qa_with_separated_facts_1 | {'train': 8134, 'validation': 926, 'test': 920} |
qasc_qa_with_separated_facts_2 | {'train': 8134, 'validation': 926, 'test': 920} |
qasc_qa_with_separated_facts_3 | {'train': 8134, 'validation': 926, 'test': 920} |
qasc_qa_with_separated_facts_4 | {'train': 8134, 'validation': 926, 'test': 920} |
qasc_qa_with_separated_facts_5 | {'train': 8134, 'validation': 926, 'test': 920} |
quail_context_description_question_answer_id | {'train': 10246, 'validation': 2164, 'challenge': 556} |
quail_context_description_question_answer_text | {'train': 10246, 'validation': 2164, 'challenge': 556} |
quail_context_description_question_text | {'train': 10246, 'validation': 2164, 'challenge': 556} |
quail_context_question_answer_description_id | {'train': 10246, 'validation': 2164, 'challenge': 556} |
quail_context_question_answer_description_text | {'train': 10246, 'validation': 2164, 'challenge': 556} |
quail_context_question_description_answer_id | {'train': 10246, 'validation': 2164, 'challenge': 556} |
quail_context_question_description_answer_text | {'train': 10246, 'validation': 2164, 'challenge': 556} |
quail_context_question_description_text | {'train': 10246, 'validation': 2164, 'challenge': 556} |
quail_description_context_question_answer_id | {'train': 10246, 'validation': 2164, 'challenge': 556} |
quail_description_context_question_answer_text | {'train': 10246, 'validation': 2164, 'challenge': 556} |
quail_description_context_question_text | {'train': 10246, 'validation': 2164, 'challenge': 556} |
quail_no_prompt_id | {'train': 10246, 'validation': 2164, 'challenge': 556} |
quail_no_prompt_text | {'train': 10246, 'validation': 2164, 'challenge': 556} |
quarel_choose_between | {'train': 1941, 'validation': 278, 'test': 552} |
quarel_do_not_use | {'train': 1941, 'validation': 278, 'test': 552} |
quarel_heres_a_story | {'train': 1941, 'validation': 278, 'test': 552} |
quarel_logic_test | {'train': 1941, 'validation': 278, 'test': 552} |
quarel_testing_students | {'train': 1941, 'validation': 278, 'test': 552} |
quartz_answer_question_based_on | {'train': 2696, 'validation': 384, 'test': 784} |
quartz_answer_question_below | {'train': 2696, 'validation': 384, 'test': 784} |
quartz_given_the_fact_answer_the_q | {'train': 2696, 'validation': 384, 'test': 784} |
quartz_having_read_above_passage | {'train': 2696, 'validation': 384, 'test': 784} |
quartz_paragraph_question_plain_concat | {'train': 2696, 'validation': 384, 'test': 784} |
quartz_read_passage_below_choose | {'train': 2696, 'validation': 384, 'test': 784} |
quartz_use_info_from_paragraph_question | {'train': 2696, 'validation': 384, 'test': 784} |
quartz_use_info_from_question_paragraph | {'train': 2696, 'validation': 384, 'test': 784} |
quoref_Answer_Friend_Question | {'train': 19399, 'validation': 2418} |
quoref_Answer_Question_Given_Context | {'train': 19399, 'validation': 2418} |
quoref_Answer_Test | {'train': 19399, 'validation': 2418} |
quoref_Context_Contains_Answer | {'train': 19399, 'validation': 2418} |
quoref_Find_Answer | {'train': 19399, 'validation': 2418} |
quoref_Found_Context_Online | {'train': 19399, 'validation': 2418} |
quoref_Given_Context_Answer_Question | {'train': 19399, 'validation': 2418} |
quoref_Guess_Answer | {'train': 19399, 'validation': 2418} |
quoref_Guess_Title_For_Context | {'train': 19399, 'validation': 2418} |
quoref_Read_And_Extract_ | {'train': 19399, 'validation': 2418} |
quoref_What_Is_The_Answer | {'train': 19399, 'validation': 2418} |
race_high_Is_this_the_right_answer | {'train': 62445, 'validation': 3451, 'test': 3498} |
race_high_Read_the_article_and_answer_the_question_no_option_ | {'train': 62445, 'validation': 3451, 'test': 3498} |
race_high_Select_the_best_answer | {'train': 62445, 'validation': 3451, 'test': 3498} |
race_high_Select_the_best_answer_generate_span_ | {'train': 62445, 'validation': 3451, 'test': 3498} |
race_high_Select_the_best_answer_no_instructions_ | {'train': 62445, 'validation': 3451, 'test': 3498} |
race_high_Taking_a_test | {'train': 62445, 'validation': 3451, 'test': 3498} |
race_high_Write_a_multi_choice_question_for_the_following_article | {'train': 62445, 'validation': 3451, 'test': 3498} |
race_high_Write_a_multi_choice_question_options_given_ | {'train': 62445, 'validation': 3451, 'test': 3498} |
race_middle_Is_this_the_right_answer | {'train': 25421, 'validation': 1436, 'test': 1436} |
race_middle_Read_the_article_and_answer_the_question_no_option_ | {'train': 25421, 'validation': 1436, 'test': 1436} |
race_middle_Select_the_best_answer | {'train': 25421, 'validation': 1436, 'test': 1436} |
race_middle_Select_the_best_answer_generate_span_ | {'train': 25421, 'validation': 1436, 'test': 1436} |
race_middle_Select_the_best_answer_no_instructions_ | {'train': 25421, 'validation': 1436, 'test': 1436} |
race_middle_Taking_a_test | {'train': 25421, 'validation': 1436, 'test': 1436} |
race_middle_Write_a_multi_choice_question_for_the_following_article | {'train': 25421, 'validation': 1436, 'test': 1436} |
race_middle_Write_a_multi_choice_question_options_given_ | {'train': 25421, 'validation': 1436, 'test': 1436} |
ropes_background_new_situation_answer | {'train': 10924, 'validation': 1688} |
ropes_background_situation_middle | {'train': 10924, 'validation': 1688} |
ropes_given_background_situation | {'train': 10924, 'validation': 1688} |
ropes_new_situation_background_answer | {'train': 10924, 'validation': 1688} |
ropes_plain_background_situation | {'train': 10924, 'validation': 1688} |
ropes_plain_bottom_hint | {'train': 10924, 'validation': 1688} |
ropes_plain_no_background | {'train': 10924, 'validation': 1688} |
ropes_prompt_beginning | {'train': 10924, 'validation': 1688} |
ropes_prompt_bottom_hint_beginning | {'train': 10924, 'validation': 1688} |
ropes_prompt_bottom_no_hint | {'train': 10924, 'validation': 1688} |
ropes_prompt_mix | {'train': 10924, 'validation': 1688} |
ropes_read_background_situation | {'train': 10924, 'validation': 1688} |
rotten_tomatoes_Movie_Expressed_Sentiment | {'train': 8530, 'validation': 1066, 'test': 1066} |
rotten_tomatoes_Movie_Expressed_Sentiment_2 | {'train': 8530, 'validation': 1066, 'test': 1066} |
rotten_tomatoes_Reviewer_Enjoyment | {'train': 8530, 'validation': 1066, 'test': 1066} |
rotten_tomatoes_Reviewer_Enjoyment_Yes_No | {'train': 8530, 'validation': 1066, 'test': 1066} |
rotten_tomatoes_Reviewer_Expressed_Sentiment | {'train': 8530, 'validation': 1066, 'test': 1066} |
rotten_tomatoes_Reviewer_Opinion_bad_good_choices | {'train': 8530, 'validation': 1066, 'test': 1066} |
rotten_tomatoes_Reviewer_Sentiment_Feeling | {'train': 8530, 'validation': 1066, 'test': 1066} |
rotten_tomatoes_Sentiment_with_choices_ | {'train': 8530, 'validation': 1066, 'test': 1066} |
rotten_tomatoes_Text_Expressed_Sentiment | {'train': 8530, 'validation': 1066, 'test': 1066} |
rotten_tomatoes_Writer_Expressed_Sentiment | {'train': 8530, 'validation': 1066, 'test': 1066} |
samsum_Generate_a_summary_for_this_dialogue | {'train': 14732, 'validation': 818, 'test': 819} |
samsum_Given_the_above_dialogue_write_a_summary | {'train': 14732, 'validation': 818, 'test': 819} |
samsum_Sum_up_the_following_dialogue | {'train': 14732, 'validation': 818, 'test': 819} |
samsum_Summarize_ | {'train': 14732, 'validation': 818, 'test': 819} |
samsum_Summarize_this_dialogue_ | {'train': 14732, 'validation': 818, 'test': 819} |
samsum_To_sum_up_this_dialog | {'train': 14732, 'validation': 818, 'test': 819} |
samsum_Write_a_dialogue_that_match_this_summary | {'train': 14732, 'validation': 818, 'test': 819} |
sciq_Direct_Question | {'train': 11679, 'validation': 1000, 'test': 1000} |
sciq_Direct_Question_Closed_Book_ | {'train': 11679, 'validation': 1000, 'test': 1000} |
sciq_Multiple_Choice | {'train': 11679, 'validation': 1000, 'test': 1000} |
sciq_Multiple_Choice_Closed_Book_ | {'train': 11679, 'validation': 1000, 'test': 1000} |
sciq_Multiple_Choice_Question_First | {'train': 11679, 'validation': 1000, 'test': 1000} |
social_i_qa_Check_if_a_random_answer_is_valid_or_not | {'train': 33410, 'validation': 1954} |
social_i_qa_Generate_answer | {'train': 33410, 'validation': 1954} |
social_i_qa_Generate_the_question_from_the_answer | {'train': 33410, 'validation': 1954} |
social_i_qa_I_was_wondering | {'train': 33410, 'validation': 1954} |
social_i_qa_Show_choices_and_generate_answer | {'train': 33410, 'validation': 1954} |
social_i_qa_Show_choices_and_generate_index | {'train': 33410, 'validation': 1954} |
squad_v2_Jeopardy_with_Context | {'train': 86821, 'validation': 5928} |
squad_v2_Jeopardy_without_Context | {'train': 86821, 'validation': 5928} |
squad_v2_Questions_with_Context | {'train': 130319, 'validation': 11873} |
squad_v2_Questions_with_Context_Without_Prompt_Keywords | {'train': 130319, 'validation': 11873} |
squad_v2_Questions_with_Context_Without_Prompt_Keywords_unanswerable | {'train': 130319, 'validation': 11873} |
squad_v2_Questions_with_Context_unanswerable | {'train': 130319, 'validation': 11873} |
squad_v2_Topic_Prediction_Context | {'train': 130319, 'validation': 11873} |
squad_v2_Topic_Prediction_Context_with_randomized_prompt_options | {'train': 130319, 'validation': 11873} |
squad_v2_Topic_Prediction_Context_with_randomized_prompt_options_placed_in_the_end | {'train': 130319, 'validation': 11873} |
squad_v2_Topic_Prediction_Question_and_Answer_Pair | {'train': 86821, 'validation': 5928} |
squad_v2_Trivia | {'train': 86821, 'validation': 5928} |
squad_v2_Unanwerable_question | {'train': 130319, 'validation': 11873} |
super_glue_boolq_GPT_3_Style | {'train': 9427, 'validation': 3270, 'test': 3245} |
super_glue_boolq_I_wonder_ | {'train': 9427, 'validation': 3270, 'test': 3245} |
super_glue_boolq_after_reading | {'train': 9427, 'validation': 3270, 'test': 3245} |
super_glue_boolq_based_on_the_following_passage | {'train': 9427, 'validation': 3270, 'test': 3245} |
super_glue_boolq_based_on_the_previous_passage | {'train': 9427, 'validation': 3270, 'test': 3245} |
super_glue_boolq_could_you_tell_me_ | {'train': 9427, 'validation': 3270, 'test': 3245} |
super_glue_boolq_exam | {'train': 9427, 'validation': 3270, 'test': 3245} |
super_glue_boolq_exercise | {'train': 9427, 'validation': 3270, 'test': 3245} |
super_glue_boolq_valid_binary | {'train': 9427, 'validation': 3270, 'test': 3245} |
super_glue_boolq_yes_no_question | {'train': 9427, 'validation': 3270, 'test': 3245} |
super_glue_cb_GPT_3_style | {'train': 250, 'validation': 56, 'test': 250} |
super_glue_cb_GPT_3_style_score_eval | {'train': 750, 'validation': 168, 'test': 750} |
super_glue_cb_MNLI_crowdsource | {'train': 250, 'validation': 56, 'test': 250} |
super_glue_cb_MNLI_crowdsource_score_eval | {'train': 750, 'validation': 168, 'test': 750} |
super_glue_cb_always_sometimes_never | {'train': 250, 'validation': 56, 'test': 250} |
super_glue_cb_always_sometimes_never_score_eval | {'train': 750, 'validation': 168, 'test': 750} |
super_glue_cb_based_on_the_previous_passage | {'train': 250, 'validation': 56, 'test': 250} |
super_glue_cb_based_on_the_previous_passage_score_eval | {'train': 750, 'validation': 168, 'test': 750} |
super_glue_cb_can_we_infer | {'train': 250, 'validation': 56, 'test': 250} |
super_glue_cb_can_we_infer_score_eval | {'train': 750, 'validation': 168, 'test': 750} |
super_glue_cb_claim_true_false_inconclusive | {'train': 250, 'validation': 56, 'test': 250} |
super_glue_cb_claim_true_false_inconclusive_score_eval | {'train': 750, 'validation': 168, 'test': 750} |
super_glue_cb_consider_always_sometimes_never | {'train': 250, 'validation': 56, 'test': 250} |
super_glue_cb_consider_always_sometimes_never_score_eval | {'train': 750, 'validation': 168, 'test': 750} |
super_glue_cb_does_it_follow_that | {'train': 250, 'validation': 56, 'test': 250} |
super_glue_cb_does_it_follow_that_score_eval | {'train': 750, 'validation': 168, 'test': 750} |
super_glue_cb_does_this_imply | {'train': 250, 'validation': 56, 'test': 250} |
super_glue_cb_does_this_imply_score_eval | {'train': 750, 'validation': 168, 'test': 750} |
super_glue_cb_guaranteed_possible_impossible | {'train': 250, 'validation': 56, 'test': 250} |
super_glue_cb_guaranteed_possible_impossible_score_eval | {'train': 750, 'validation': 168, 'test': 750} |
super_glue_cb_guaranteed_true | {'train': 250, 'validation': 56, 'test': 250} |
super_glue_cb_guaranteed_true_score_eval | {'train': 750, 'validation': 168, 'test': 750} |
super_glue_cb_justified_in_saying | {'train': 250, 'validation': 56, 'test': 250} |
super_glue_cb_justified_in_saying_score_eval | {'train': 750, 'validation': 168, 'test': 750} |
super_glue_cb_must_be_true | {'train': 250, 'validation': 56, 'test': 250} |
super_glue_cb_must_be_true_score_eval | {'train': 750, 'validation': 168, 'test': 750} |
super_glue_cb_should_assume | {'train': 250, 'validation': 56, 'test': 250} |
super_glue_cb_should_assume_score_eval | {'train': 750, 'validation': 168, 'test': 750} |
super_glue_cb_take_the_following_as_truth | {'train': 250, 'validation': 56, 'test': 250} |
super_glue_cb_take_the_following_as_truth_score_eval | {'train': 750, 'validation': 168, 'test': 750} |
super_glue_copa_C1_or_C2_premise_so_because_ | {'train': 400, 'validation': 100, 'test': 500} |
super_glue_copa_C1_or_C2_premise_so_because__score_eval | {'train': 800, 'validation': 200, 'test': 1000} |
super_glue_copa__As_a_result_C1_or_C2_ | {'train': 202, 'validation': 48, 'test': 250} |
super_glue_copa__As_a_result_C1_or_C2__score_eval | {'train': 404, 'validation': 96, 'test': 500} |
super_glue_copa__What_could_happen_next_C1_or_C2_ | {'train': 202, 'validation': 48, 'test': 250} |
super_glue_copa__What_could_happen_next_C1_or_C2__score_eval | {'train': 404, 'validation': 96, 'test': 500} |
super_glue_copa__which_may_be_caused_by | {'train': 198, 'validation': 52, 'test': 250} |
super_glue_copa__which_may_be_caused_by_score_eval | {'train': 396, 'validation': 104, 'test': 500} |
super_glue_copa__why_C1_or_C2 | {'train': 198, 'validation': 52, 'test': 250} |
super_glue_copa__why_C1_or_C2_score_eval | {'train': 396, 'validation': 104, 'test': 500} |
super_glue_copa_best_option | {'train': 400, 'validation': 100, 'test': 500} |
super_glue_copa_best_option_score_eval | {'train': 800, 'validation': 200, 'test': 1000} |
super_glue_copa_cause_effect | {'train': 400, 'validation': 100, 'test': 500} |
super_glue_copa_cause_effect_score_eval | {'train': 800, 'validation': 200, 'test': 1000} |
super_glue_copa_choose | {'train': 400, 'validation': 100, 'test': 500} |
super_glue_copa_choose_score_eval | {'train': 800, 'validation': 200, 'test': 1000} |
super_glue_copa_exercise | {'train': 400, 'validation': 100, 'test': 500} |
super_glue_copa_exercise_score_eval | {'train': 800, 'validation': 200, 'test': 1000} |
super_glue_copa_i_am_hesitating | {'train': 400, 'validation': 100, 'test': 500} |
super_glue_copa_i_am_hesitating_score_eval | {'train': 800, 'validation': 200, 'test': 1000} |
super_glue_copa_more_likely | {'train': 400, 'validation': 100, 'test': 500} |
super_glue_copa_more_likely_score_eval | {'train': 800, 'validation': 200, 'test': 1000} |
super_glue_copa_plausible_alternatives | {'train': 400, 'validation': 100, 'test': 500} |
super_glue_copa_plausible_alternatives_score_eval | {'train': 800, 'validation': 200, 'test': 1000} |
super_glue_multirc_I_was_going_to_say_ | {'train': 27243, 'validation': 4848, 'test': 9693} |
super_glue_multirc_Would_it_be_good_to_answer_ | {'train': 27243, 'validation': 4848, 'test': 9693} |
super_glue_multirc_confirm | {'train': 27243, 'validation': 4848, 'test': 9693} |
super_glue_multirc_correct | {'train': 27243, 'validation': 4848, 'test': 9693} |
super_glue_multirc_decide_valid | {'train': 27243, 'validation': 4848, 'test': 9693} |
super_glue_multirc_found_this_answer | {'train': 27243, 'validation': 4848, 'test': 9693} |
super_glue_multirc_grading | {'train': 27243, 'validation': 4848, 'test': 9693} |
super_glue_multirc_is_a_correct_answer_ | {'train': 27243, 'validation': 4848, 'test': 9693} |
super_glue_multirc_is_the_correct_answer_ | {'train': 27243, 'validation': 4848, 'test': 9693} |
super_glue_multirc_paragraph_question_is_it_ | {'train': 27243, 'validation': 4848, 'test': 9693} |
super_glue_record_Add_sentence_after_after_continuation_choices_ | {'train': 100730, 'validation': 10000, 'test': 10000} |
super_glue_record_Add_sentence_after_continuation_choices_ | {'train': 100730, 'validation': 10000, 'test': 10000} |
super_glue_record_Can_you_figure_out_ | {'train': 100730, 'validation': 10000, 'test': 10000} |
super_glue_record_GPT_3_style_continuation_choices_ | {'train': 100730, 'validation': 10000, 'test': 10000} |
super_glue_record_GPT_3_style_summary_only_continuation_choices_ | {'train': 100730, 'validation': 10000, 'test': 10000} |
super_glue_record_GPT_3_style_with_labels_continuation_choices_ | {'train': 100730, 'validation': 10000, 'test': 10000} |
super_glue_record_GPT_3_style_with_labels_without_hyphens_continuation_choices_ | {'train': 100730, 'validation': 10000, 'test': 10000} |
super_glue_record_GPT_3_style_without_hyphens_continuation_choices_ | {'train': 100730, 'validation': 10000, 'test': 10000} |
super_glue_record_In_the_question_above_the_placeholder_stands_for | {'train': 100730, 'validation': 10000, 'test': 10000} |
super_glue_record_New_highlight_continuation_choices_ | {'train': 100730, 'validation': 10000, 'test': 10000} |
super_glue_record_News_article_continuation_choices_ | {'train': 100730, 'validation': 10000, 'test': 10000} |
super_glue_record_Summary_first_continuation_choices_ | {'train': 100730, 'validation': 10000, 'test': 10000} |
super_glue_record_What_could_the_placeholder_be_ | {'train': 100730, 'validation': 10000, 'test': 10000} |
super_glue_record_Which_one_is_the_placeholder_ | {'train': 100730, 'validation': 10000, 'test': 10000} |
super_glue_record_choose_between | {'train': 100730, 'validation': 10000, 'test': 10000} |
super_glue_record_corrupted | {'train': 100730, 'validation': 10000, 'test': 10000} |
super_glue_record_exercise | {'train': 100730, 'validation': 10000, 'test': 10000} |
super_glue_record_pick_one_option | {'train': 100730, 'validation': 10000, 'test': 10000} |
super_glue_record_the_placeholder_refers_to_ | {'train': 100730, 'validation': 10000, 'test': 10000} |
super_glue_record_trying_to_decide | {'train': 100730, 'validation': 10000, 'test': 10000} |
super_glue_rte_GPT_3_style | {'train': 2490, 'validation': 277, 'test': 3000} |
super_glue_rte_GPT_3_style_score_eval | {'train': 4980, 'validation': 554, 'test': 6000} |
super_glue_rte_MNLI_crowdsource | {'train': 2490, 'validation': 277, 'test': 3000} |
super_glue_rte_MNLI_crowdsource_score_eval | {'train': 4980, 'validation': 554, 'test': 6000} |
super_glue_rte_based_on_the_previous_passage | {'train': 2490, 'validation': 277, 'test': 3000} |
super_glue_rte_based_on_the_previous_passage_score_eval | {'train': 4980, 'validation': 554, 'test': 6000} |
super_glue_rte_can_we_infer | {'train': 2490, 'validation': 277, 'test': 3000} |
super_glue_rte_can_we_infer_score_eval | {'train': 4980, 'validation': 554, 'test': 6000} |
super_glue_rte_does_it_follow_that | {'train': 2490, 'validation': 277, 'test': 3000} |
super_glue_rte_does_it_follow_that_score_eval | {'train': 4980, 'validation': 554, 'test': 6000} |
super_glue_rte_does_this_imply | {'train': 2490, 'validation': 277, 'test': 3000} |
super_glue_rte_does_this_imply_score_eval | {'train': 4980, 'validation': 554, 'test': 6000} |
super_glue_rte_guaranteed_true | {'train': 2490, 'validation': 277, 'test': 3000} |
super_glue_rte_guaranteed_true_score_eval | {'train': 4980, 'validation': 554, 'test': 6000} |
super_glue_rte_justified_in_saying | {'train': 2490, 'validation': 277, 'test': 3000} |
super_glue_rte_justified_in_saying_score_eval | {'train': 4980, 'validation': 554, 'test': 6000} |
super_glue_rte_must_be_true | {'train': 2490, 'validation': 277, 'test': 3000} |
super_glue_rte_must_be_true_score_eval | {'train': 4980, 'validation': 554, 'test': 6000} |
super_glue_rte_should_assume | {'train': 2490, 'validation': 277, 'test': 3000} |
super_glue_rte_should_assume_score_eval | {'train': 4980, 'validation': 554, 'test': 6000} |
super_glue_wic_GPT_3_prompt | {'train': 5428, 'validation': 638, 'test': 1400} |
super_glue_wic_GPT_3_prompt_score_eval | {'train': 10856, 'validation': 1276, 'test': 2800} |
super_glue_wic_GPT_3_prompt_with_label | {'train': 5428, 'validation': 638, 'test': 1400} |
super_glue_wic_GPT_3_prompt_with_label_score_eval | {'train': 10856, 'validation': 1276, 'test': 2800} |
super_glue_wic_affirmation_true_or_false | {'train': 5428, 'validation': 638, 'test': 1400} |
super_glue_wic_affirmation_true_or_false_score_eval | {'train': 10856, 'validation': 1276, 'test': 2800} |
super_glue_wic_grammar_homework | {'train': 5428, 'validation': 638, 'test': 1400} |
super_glue_wic_grammar_homework_score_eval | {'train': 10856, 'validation': 1276, 'test': 2800} |
super_glue_wic_polysemous | {'train': 5428, 'validation': 638, 'test': 1400} |
super_glue_wic_polysemous_score_eval | {'train': 10856, 'validation': 1276, 'test': 2800} |
super_glue_wic_question_context | {'train': 5428, 'validation': 638, 'test': 1400} |
super_glue_wic_question_context_meaning | {'train': 5428, 'validation': 638, 'test': 1400} |
super_glue_wic_question_context_meaning_score_eval | {'train': 10856, 'validation': 1276, 'test': 2800} |
super_glue_wic_question_context_meaning_with_label | {'train': 5428, 'validation': 638, 'test': 1400} |
super_glue_wic_question_context_meaning_with_label_score_eval | {'train': 10856, 'validation': 1276, 'test': 2800} |
super_glue_wic_question_context_score_eval | {'train': 10856, 'validation': 1276, 'test': 2800} |
super_glue_wic_same_sense | {'train': 5428, 'validation': 638, 'test': 1400} |
super_glue_wic_same_sense_score_eval | {'train': 10856, 'validation': 1276, 'test': 2800} |
super_glue_wic_similar_sense | {'train': 5428, 'validation': 638, 'test': 1400} |
super_glue_wic_similar_sense_score_eval | {'train': 10856, 'validation': 1276, 'test': 2800} |
super_glue_wsc.fixed_GPT_3_Style | {'train': 554, 'validation': 104, 'test': 146} |
super_glue_wsc.fixed_GPT_3_Style_score_eval | {'train': 1108, 'validation': 208, 'test': 292} |
super_glue_wsc.fixed_I_think_they_mean | {'train': 554, 'validation': 104, 'test': 146} |
super_glue_wsc.fixed_I_think_they_mean_score_eval | {'train': 1108, 'validation': 208, 'test': 292} |
super_glue_wsc.fixed_Who_or_what_is_are | {'train': 554, 'validation': 104, 'test': 146} |
super_glue_wsc.fixed_Who_or_what_is_are_score_eval | {'train': 1108, 'validation': 208, 'test': 292} |
super_glue_wsc.fixed_by_p_they_mean | {'train': 554, 'validation': 104, 'test': 146} |
super_glue_wsc.fixed_by_p_they_mean_score_eval | {'train': 1108, 'validation': 208, 'test': 292} |
super_glue_wsc.fixed_does_p_stand_for | {'train': 554, 'validation': 104, 'test': 146} |
super_glue_wsc.fixed_does_p_stand_for_score_eval | {'train': 1108, 'validation': 208, 'test': 292} |
super_glue_wsc.fixed_does_the_pronoun_refer_to | {'train': 554, 'validation': 104, 'test': 146} |
super_glue_wsc.fixed_does_the_pronoun_refer_to_score_eval | {'train': 1108, 'validation': 208, 'test': 292} |
super_glue_wsc.fixed_in_other_words | {'train': 554, 'validation': 104, 'test': 146} |
super_glue_wsc.fixed_in_other_words_score_eval | {'train': 1108, 'validation': 208, 'test': 292} |
super_glue_wsc.fixed_p_is_are_r | {'train': 554, 'validation': 104, 'test': 146} |
super_glue_wsc.fixed_p_is_are_r_score_eval | {'train': 1108, 'validation': 208, 'test': 292} |
super_glue_wsc.fixed_replaced_with | {'train': 554, 'validation': 104, 'test': 146} |
super_glue_wsc.fixed_replaced_with_score_eval | {'train': 1108, 'validation': 208, 'test': 292} |
super_glue_wsc.fixed_the_pronoun_refers_to | {'train': 554, 'validation': 104, 'test': 146} |
super_glue_wsc.fixed_the_pronoun_refers_to_score_eval | {'train': 1108, 'validation': 208, 'test': 292} |
trec_fine_grained_ABBR | {'train': 86, 'test': 9} |
trec_fine_grained_ABBR_context_first | {'train': 86, 'test': 9} |
trec_fine_grained_DESC | {'train': 1162, 'test': 138} |
trec_fine_grained_DESC_context_first | {'train': 1162, 'test': 138} |
trec_fine_grained_ENTY | {'train': 1250, 'test': 94} |
trec_fine_grained_HUM | {'train': 1223, 'test': 65} |
trec_fine_grained_HUM_context_first | {'train': 1223, 'test': 65} |
trec_fine_grained_LOC | {'train': 835, 'test': 81} |
trec_fine_grained_LOC_context_first | {'train': 835, 'test': 81} |
trec_fine_grained_NUM | {'train': 896, 'test': 113} |
trec_fine_grained_NUM_context_first | {'train': 896, 'test': 113} |
trec_fine_grained_open | {'train': 5452, 'test': 500} |
trec_fine_grained_open_context_first | {'train': 5452, 'test': 500} |
trec_pick_the_best_descriptor | {'train': 5452, 'test': 500} |
trec_trec1 | {'train': 5452, 'test': 500} |
trec_trec2 | {'train': 5452, 'test': 500} |
trec_what_category_best_describe | {'train': 5452, 'test': 500} |
trec_which_category_best_describes | {'train': 5452, 'test': 500} |
trivia_qa_unfiltered_first_person_context | {'train': 87622, 'validation': 11313, 'test': 10832} |
trivia_qa_unfiltered_formal_description | {'train': 87622, 'validation': 11313, 'test': 10832} |
trivia_qa_unfiltered_guess_question | {'train': 87622, 'validation': 11313} |
trivia_qa_unfiltered_question_answer | {'train': 87622, 'validation': 11313, 'test': 10832} |
trivia_qa_unfiltered_question_with_instruction | {'train': 87622, 'validation': 11313, 'test': 10832} |
web_questions_get_the_answer | {'train': 3778, 'test': 2032} |
web_questions_potential_correct_answer | {'train': 3778, 'test': 2032} |
web_questions_question_answer | {'train': 3778, 'test': 2032} |
web_questions_short_general_knowledge_q | {'train': 3778, 'test': 2032} |
web_questions_whats_the_answer | {'train': 3778, 'test': 2032} |
wiki_bio_comprehension | {'train': 582639, 'test': 72829, 'val': 72831} |
wiki_bio_guess_person | {'train': 582639, 'test': 72829, 'val': 72831} |
wiki_bio_key_content | {'train': 582639, 'test': 72829, 'val': 72831} |
wiki_bio_what_content | {'train': 582639, 'test': 72829, 'val': 72831} |
wiki_bio_who | {'train': 582639, 'test': 72829, 'val': 72831} |
wiki_hop_original_choose_best_object_affirmative_1 | {'train': 43738, 'validation': 5129} |
wiki_hop_original_choose_best_object_affirmative_2 | {'train': 43738, 'validation': 5129} |
wiki_hop_original_choose_best_object_affirmative_3 | {'train': 43738, 'validation': 5129} |
wiki_hop_original_choose_best_object_interrogative_1 | {'train': 43738, 'validation': 5129} |
wiki_hop_original_choose_best_object_interrogative_2 | {'train': 43738, 'validation': 5129} |
wiki_hop_original_explain_relation | {'train': 43738, 'validation': 5129} |
wiki_hop_original_generate_object | {'train': 43738, 'validation': 5129} |
wiki_hop_original_generate_subject | {'train': 43738, 'validation': 5129} |
wiki_hop_original_generate_subject_and_object | {'train': 43738, 'validation': 5129} |
wiki_qa_Decide_good_answer | {'train': 20360, 'validation': 2733, 'test': 6165} |
wiki_qa_Direct_Answer_to_Question | {'train': 1040, 'validation': 140, 'test': 293} |
wiki_qa_Generate_Question_from_Topic | {'train': 1040, 'validation': 140, 'test': 293} |
wiki_qa_Is_This_True_ | {'train': 20360, 'validation': 2733, 'test': 6165} |
wiki_qa_Jeopardy_style | {'train': 1040, 'validation': 140, 'test': 293} |
wiki_qa_Topic_Prediction_Answer_Only | {'train': 1040, 'validation': 140, 'test': 293} |
wiki_qa_Topic_Prediction_Question_Only | {'train': 1040, 'validation': 140, 'test': 293} |
wiki_qa_Topic_Prediction_Question_and_Answer_Pair | {'train': 1040, 'validation': 140, 'test': 293} |
wiki_qa_automatic_system | {'train': 20360, 'validation': 2733, 'test': 6165} |
wiki_qa_exercise | {'train': 20360, 'validation': 2733, 'test': 6165} |
wiki_qa_found_on_google | {'train': 20360, 'validation': 2733, 'test': 6165} |
winogrande_winogrande_debiased_Replace | {'train': 9248, 'validation': 1267, 'test': 1767} |
winogrande_winogrande_debiased_Replace_score_eval | {'train': 18496, 'validation': 2534, 'test': 3534} |
winogrande_winogrande_debiased_does_underscore_refer_to | {'train': 9248, 'validation': 1267, 'test': 1767} |
winogrande_winogrande_debiased_does_underscore_refer_to_score_eval | {'train': 18496, 'validation': 2534, 'test': 3534} |
winogrande_winogrande_debiased_fill_in_the_blank | {'train': 9248, 'validation': 1267, 'test': 1767} |
winogrande_winogrande_debiased_fill_in_the_blank_score_eval | {'train': 18496, 'validation': 2534, 'test': 3534} |
winogrande_winogrande_debiased_stand_for | {'train': 9248, 'validation': 1267, 'test': 1767} |
winogrande_winogrande_debiased_stand_for_score_eval | {'train': 18496, 'validation': 2534, 'test': 3534} |
winogrande_winogrande_debiased_underscore_refer_to | {'train': 9248, 'validation': 1267, 'test': 1767} |
winogrande_winogrande_debiased_underscore_refer_to_score_eval | {'train': 18496, 'validation': 2534, 'test': 3534} |
winogrande_winogrande_xl_Replace | {'train': 40398, 'validation': 1267, 'test': 1767} |
winogrande_winogrande_xl_Replace_score_eval | {'train': 80796, 'validation': 2534, 'test': 3534} |
winogrande_winogrande_xl_does_underscore_refer_to | {'train': 40398, 'validation': 1267, 'test': 1767} |
winogrande_winogrande_xl_does_underscore_refer_to_score_eval | {'train': 80796, 'validation': 2534, 'test': 3534} |
winogrande_winogrande_xl_fill_in_the_blank | {'train': 40398, 'validation': 1267, 'test': 1767} |
winogrande_winogrande_xl_fill_in_the_blank_score_eval | {'train': 80796, 'validation': 2534, 'test': 3534} |
winogrande_winogrande_xl_stand_for | {'train': 40398, 'validation': 1267, 'test': 1767} |
winogrande_winogrande_xl_stand_for_score_eval | {'train': 80796, 'validation': 2534, 'test': 3534} |
winogrande_winogrande_xl_underscore_refer_to | {'train': 40398, 'validation': 1267, 'test': 1767} |
winogrande_winogrande_xl_underscore_refer_to_score_eval | {'train': 80796, 'validation': 2534, 'test': 3534} |
wiqa_does_the_supposed_perturbation_have_an_effect | {'train': 29808, 'validation': 6894, 'test': 3003} |
wiqa_effect_with_label_answer | {'train': 29808, 'validation': 6894, 'test': 3003} |
wiqa_effect_with_string_answer | {'train': 29808, 'validation': 6894, 'test': 3003} |
wiqa_what_is_the_final_step_of_the_following_process | {'train': 29808, 'validation': 6894, 'test': 3003} |
wiqa_what_is_the_missing_first_step | {'train': 29808, 'validation': 6894, 'test': 3003} |
wiqa_what_might_be_the_first_step_of_the_process | {'train': 29808, 'validation': 6894, 'test': 3003} |
wiqa_what_might_be_the_last_step_of_the_process | {'train': 29808, 'validation': 6894, 'test': 3003} |
wiqa_which_of_the_following_is_the_supposed_perturbation | {'train': 29808, 'validation': 6894, 'test': 3003} |
xsum_DOC_boils_down_to_simple_idea_that | {'train': 204045, 'validation': 11332, 'test': 11334} |
xsum_DOC_given_above_write_one_sentence | {'train': 204045, 'validation': 11332, 'test': 11334} |
xsum_DOC_how_would_you_rephrase_few_words | {'train': 204045, 'validation': 11332, 'test': 11334} |
xsum_DOC_tldr | {'train': 204045, 'validation': 11332, 'test': 11334} |
xsum_DOC_write_summary_of_above | {'train': 204045, 'validation': 11332, 'test': 11334} |
xsum_article_DOC_summary | {'train': 204045, 'validation': 11332, 'test': 11334} |
xsum_college_roommate_asked_DOC_so_I_recap | {'train': 204045, 'validation': 11332, 'test': 11334} |
xsum_read_below_DOC_write_abstract | {'train': 204045, 'validation': 11332, 'test': 11334} |
xsum_summarize_DOC | {'train': 204045, 'validation': 11332, 'test': 11334} |
xsum_summarize_this_DOC_summary | {'train': 204045, 'validation': 11332, 'test': 11334} |
yelp_review_full_based_on_that | {'train': 650000, 'test': 50000} |
yelp_review_full_format_rating | {'train': 650000, 'test': 50000} |
yelp_review_full_format_score | {'train': 650000, 'test': 50000} |
yelp_review_full_format_star | {'train': 650000, 'test': 50000} |
yelp_review_full_on_a_scale | {'train': 650000, 'test': 50000} |
yelp_review_full_so_i_would | {'train': 650000, 'test': 50000} |
yelp_review_full_this_place | {'train': 650000, 'test': 50000} |
Dataset Creation
Curation Rationale
The Public Pool of Prompts relies on the Hugging Face Dataset library. Any public dataset in the Datasets library can be prompted. We select the datasets that have at least one subset in English and excluded datasets containing (predominantly) non-natural language examples.
We conservatively decided not to prompt datasets that contain potentially harmful content (for instance, datasets built on social media content). However, we sometimes prompt datasets that are purposefully built to measure bias and fairness of trained models, and reserve these prompted datasets (the validation or test sets) for evaluation purposes.
Source Data
Here's the full list of the datasets present in the materialized version of P3:
- Multiple-Choice QA
- CommonsenseQA
- DREAM
- QUAIL
- QuaRTz
- Social IQA
- WiQA
- Cosmos
- QASC
- Quarel
- SciQ
- Wiki Hop
- ARC
- OpenBookQA
- MultiRC
- PIQA
- RACE
- HellaSwag
- BoolQ
- Extractive QA
- Adversarial QA
- Quoref
- DuoRC
- ROPES
- SQuAD v2
- ReCoRD
- Close-book QA
- Hotpot QA
- Wiki QA
- Trivia QA
- Web Questions
- Structure-to-text
- Common Gen
- Wiki Bio
- Sentiment
- Amazon
- App Reviews
- IMDB
- Rotten Tomatoes
- Yelp
- Summarization
- CNN Daily Mail
- Gigaword
- MultiNews
- SamSum
- XSum
- Topic Classification
- AG News
- DBPedia
- TREC
- Paraphrase Identification
- MRPC
- PAWS
- QQP
- Natural Language Inference
- ANLI
- CB
- RTE
- Coreference Resolution
- WSC
- Winogrande
- Word Sense disambiguation
- WiC
- Sentence Completion
- COPA
- HellaSwag
- Story Cloze
Annotations
The prompts available in Promptsource are collected as part of BigScience, one-year long research workshop on large multilingual models and datasets. 36 contributors affiliated with 24 institutions in 8 countries participated to the prompt collection. Contributors are in majority machine learning researchers or machine learning engineers.
The main annotation guideline was that prompts needed to be grammatical and understandable by a native English speaker with no prior experience of the tasks. Additionally, prompts that required explicit counting or numerical indexing were removed in favor of natural language variants, e.g., instead of predicting indices of a span to extract (e.g. in extractive question answering), the model was expected to copy the span's text instead. With these minimal constraints, prompt writers were encouraged to use both formal and creative prompts and various orderings of the data. Most of the prompts correspond directly to a version of the original proposed task, although we also allowed prompts that permuted the original task (for instance, generating a document from its summary) or allowed for ambiguous output (for instance, not indicating a list of available choices).
The full annotation given to the contributors can be found here. *Note to self: the link is currently being updated with the)
Additional Information
Licensing Information
The dataset is released under Apache 2.0.
Citation Information
TODO
Contributions
Thanks to the contributors of promptsource for adding this dataset.