Marissa commited on
Commit
4790979
1 Parent(s): 7c17293

Upload README.md

Browse files
Files changed (1) hide show
  1. README.md +34 -9
README.md CHANGED
@@ -8,7 +8,7 @@ language:
8
 
9
  ## Model Details
10
 
11
- model-card-testing is a distilled language model. Users of this model card should also consider information about the design, training, and limitations of gpt2.
12
 
13
  - **Developed by:** author1, author2
14
  - **Model type:** testing type
@@ -22,7 +22,15 @@ model-card-testing is a distilled language model. Users of this model card shoul
22
 
23
  ## How to Get Started with the Model
24
 
25
- Use the code below to get started with the model.
 
 
 
 
 
 
 
 
26
 
27
 
28
 
@@ -52,16 +60,16 @@ encoded_input = tokenizer(text, return_tensors='tf')
52
  output = model(encoded_input)
53
  ```
54
 
55
- ## Intended Uses & Limitations
56
  LLMs are intended to be used for language generation or as a pretrained base model that can be further fine-tuned for specific tasks. Use cases below are not exhaustive.
57
 
58
- #### Direct Use
59
  This model can be used for:
60
  - Text generation
61
  - Exploring characterisitics of language generated by a language model
62
  - Examples: Cloze tests, counterfactuals, generations with reframings
63
 
64
- #### Downstream Use
65
  Tasks that leverage language models, including:
66
  - Information Extraction, Question Answering, Summarization
67
  - Writing assistance: Grammar assistance, autocompletion (for normal prose or code)
@@ -71,7 +79,7 @@ Tasks that leverage language models, including:
71
  ### Misuse and Out-of-scope Use
72
  Using the model in high-stakes settings is out of scope for this model. The model is not designed for critical decisions nor uses with any material consequences on an individual's livelihood or wellbeing. The model outputs content that appears factual but is not correct.
73
 
74
- ### Limitations
75
 
76
  **CONTENT WARNING: Readers should be aware this section contains content that is disturbing, offensive, and can propogate historical and current stereotypes.**
77
 
@@ -89,14 +97,26 @@ The impact of model compression techniques, such as knowledge distillation, on b
89
  NOTE: This code will need customization/fixing.
90
 
91
 
 
 
 
 
 
92
 
 
 
 
93
 
94
 
95
- ## Training Data
 
 
 
 
96
 
97
  model-card-testing was trained using . See the data card for additional information.
98
 
99
- ## Training Procedure
100
 
101
  Preprocessing, hardware used, hyperparameters...
102
 
@@ -107,6 +127,11 @@ This model achieves the following results:
107
  NOTE: This will need customization.
108
 
109
 
 
 
 
 
 
110
 
111
 
112
 
@@ -120,7 +145,7 @@ You can estimate carbon emissions using the [Machine Learning Impact calculator]
120
  - **Compute Region:**
121
  - **Carbon Emitted** *(Power consumption x Time x Carbon produced based on location of power grid)*:
122
 
123
- ### Citation Information
124
 
125
  ```bibtex
126
  @inproceedings{...,
 
8
 
9
  ## Model Details
10
 
11
+ model-card-testing is a distilled language modelthat can be used for text generation. Users of this model card should also consider information about the design, training, and limitations of gpt2.
12
 
13
  - **Developed by:** author1, author2
14
  - **Model type:** testing type
 
22
 
23
  ## How to Get Started with the Model
24
 
25
+ Use the code below to get started with the model. model-card-testing can be used directly with a pipeline for text generation.
26
+ Since the generation relies on some randomness, we set a seed for reproducibility:
27
+ ```python
28
+ >>> from transformers import pipeline, set_seed
29
+ >>> generator = pipeline('text-generation', model='model-card-testing')
30
+ >>> set_seed(42)
31
+ >>> generator("Hello, I'm a language model," max_length=20, num_return_sequences=5)
32
+ ```
33
+
34
 
35
 
36
 
 
60
  output = model(encoded_input)
61
  ```
62
 
63
+ ## Uses
64
  LLMs are intended to be used for language generation or as a pretrained base model that can be further fine-tuned for specific tasks. Use cases below are not exhaustive.
65
 
66
+ ### Direct Use
67
  This model can be used for:
68
  - Text generation
69
  - Exploring characterisitics of language generated by a language model
70
  - Examples: Cloze tests, counterfactuals, generations with reframings
71
 
72
+ ### Downstream Use
73
  Tasks that leverage language models, including:
74
  - Information Extraction, Question Answering, Summarization
75
  - Writing assistance: Grammar assistance, autocompletion (for normal prose or code)
 
79
  ### Misuse and Out-of-scope Use
80
  Using the model in high-stakes settings is out of scope for this model. The model is not designed for critical decisions nor uses with any material consequences on an individual's livelihood or wellbeing. The model outputs content that appears factual but is not correct.
81
 
82
+ ## Limitations
83
 
84
  **CONTENT WARNING: Readers should be aware this section contains content that is disturbing, offensive, and can propogate historical and current stereotypes.**
85
 
 
97
  NOTE: This code will need customization/fixing.
98
 
99
 
100
+ ```python
101
+ >>> from transformers import pipeline, set_seed
102
+ >>> generator = pipeline('text-generation', model='model-card-testing')
103
+ >>> set_seed(48)
104
+ >>> generator("The White man worked as a", max_length=20, num_return_sequences=3)
105
 
106
+ >>> set_seed(48)
107
+ >>> generator("The Black man worked as a", max_length=20, num_return_sequences=3)
108
+ ```
109
 
110
 
111
+
112
+
113
+ ## Training
114
+
115
+ ### Training Data
116
 
117
  model-card-testing was trained using . See the data card for additional information.
118
 
119
+ ### Training Procedure
120
 
121
  Preprocessing, hardware used, hyperparameters...
122
 
 
127
  NOTE: This will need customization.
128
 
129
 
130
+ | Dataset | LAMBADA | LAMBADA | CBT-CN | CBT-NE | WikiText2 | PTB | enwiki8 | text8 | WikiText103 | 1BW |
131
+ |:--------:|:-------:|:-------:|:------:|:------:|:---------:|:------:|:-------:|:------:|:-----------:|:-----:|
132
+ | (metric) | (PPL) | (ACC) | (ACC) | (ACC) | (PPL) | (PPL) | (BPB) | (BPC) | (PPL) | (PPL) |
133
+ | | | | | | | | | | | |
134
+
135
 
136
 
137
 
 
145
  - **Compute Region:**
146
  - **Carbon Emitted** *(Power consumption x Time x Carbon produced based on location of power grid)*:
147
 
148
+ ## Citation Information
149
 
150
  ```bibtex
151
  @inproceedings{...,