chiyuzhang commited on
Commit
7256c50
1 Parent(s): ddd6b0a

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +11 -10
README.md CHANGED
@@ -2,31 +2,26 @@
2
  license: apache-2.0
3
  tags:
4
  - generated_from_trainer
 
5
  model-index:
6
  - name: flan-t5-small-distil-v2
7
  results: []
8
  language:
9
  - en
 
10
  ---
11
 
12
  <!-- This model card has been generated automatically according to the information the Trainer had access to. You
13
  should probably proofread and complete it, then remove this comment. -->
14
 
15
- # flan-t5-small-distil-v2
16
 
17
- This model is a fine-tuned version of [google/flan-t5-small](https://huggingface.co/google/flan-t5-small) on an unknown dataset.
18
 
19
  ## Model description
20
 
21
- More information needed
22
-
23
- ## Intended uses & limitations
24
 
25
- More information needed
26
-
27
- ## Training and evaluation data
28
-
29
- More information needed
30
 
31
  ## Training procedure
32
 
@@ -43,6 +38,10 @@ The following hyperparameters were used during training:
43
  - lr_scheduler_type: linear
44
  - num_epochs: 5
45
 
 
 
 
 
46
  ## Use
47
 
48
  ### CPU
@@ -87,7 +86,9 @@ print("Response": generated_text)
87
 
88
  </details>
89
 
 
90
 
 
91
 
92
 
93
  ### Framework versions
 
2
  license: apache-2.0
3
  tags:
4
  - generated_from_trainer
5
+ - instruction fine-tuning
6
  model-index:
7
  - name: flan-t5-small-distil-v2
8
  results: []
9
  language:
10
  - en
11
+ pipeline_tag: text2text-generation
12
  ---
13
 
14
  <!-- This model card has been generated automatically according to the information the Trainer had access to. You
15
  should probably proofread and complete it, then remove this comment. -->
16
 
17
+ # LaMini-FLAN-T5-Small
18
 
19
+ This model is a fine-tuned version of [google/flan-t5-small](https://huggingface.co/google/flan-t5-small) on [LaMini dataset]() that contains 2.58M samples for instruction fine-tuning. For more information about our dataset, please refer to our [project repository]().
20
 
21
  ## Model description
22
 
23
+ We initialize with [google/flan-t5-small](https://huggingface.co/google/flan-t5-small) and fine-tune it on our [LaMini dataset](). Its total number of parameters is 61M.
 
 
24
 
 
 
 
 
 
25
 
26
  ## Training procedure
27
 
 
38
  - lr_scheduler_type: linear
39
  - num_epochs: 5
40
 
41
+ ## Training and evaluation data
42
+ We conducted two sets of evaluations: automatic evaluation on downstream NLP tasks and human evaluation on user-oriented instructions. For more detail, please refer to our [paper]().
43
+
44
+
45
  ## Use
46
 
47
  ### CPU
 
86
 
87
  </details>
88
 
89
+ ## Intended uses & limitations
90
 
91
+ More information needed
92
 
93
 
94
  ### Framework versions