alonzogarbanzo commited on
Commit
01995b2
·
verified ·
1 Parent(s): 7a88c71

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +12 -6
README.md CHANGED
@@ -8,21 +8,17 @@ model-index:
8
  results: []
9
  ---
10
 
11
- <!-- This model card has been generated automatically according to the information the Trainer had access to. You
12
- should probably proofread and complete it, then remove this comment. -->
13
-
14
  # Bloom-1b7-creative-writing
15
 
16
  This model is a fine-tuned version of [bigscience/bloom-1b7](https://huggingface.co/bigscience/bloom-1b7) on the [adambjorn/UnrelatedForgettingOverhead](https://huggingface.co/datasets/adambjorn/UnrelatedForgettingOverhead) creative writing dataset.
17
 
18
-
19
  ## Model description
20
 
21
  More information needed
22
 
23
  ## Intended uses & limitations
24
 
25
- More information needed
26
 
27
  ## Training and evaluation data
28
 
@@ -30,6 +26,8 @@ More information needed
30
 
31
  ## Training procedure
32
 
 
 
33
  ### Training hyperparameters
34
 
35
  The following hyperparameters were used during training:
@@ -45,7 +43,15 @@ The following hyperparameters were used during training:
45
  - mixed_precision_training: Native AMP
46
 
47
  ### Training results
48
-
 
 
 
 
 
 
 
 
49
 
50
 
51
  ### Framework versions
 
8
  results: []
9
  ---
10
 
 
 
 
11
  # Bloom-1b7-creative-writing
12
 
13
  This model is a fine-tuned version of [bigscience/bloom-1b7](https://huggingface.co/bigscience/bloom-1b7) on the [adambjorn/UnrelatedForgettingOverhead](https://huggingface.co/datasets/adambjorn/UnrelatedForgettingOverhead) creative writing dataset.
14
 
 
15
  ## Model description
16
 
17
  More information needed
18
 
19
  ## Intended uses & limitations
20
 
21
+ Intended for use on a student group project for Portland State University's Winter 2024 LLMs Course.
22
 
23
  ## Training and evaluation data
24
 
 
26
 
27
  ## Training procedure
28
 
29
+ Trained on a single RTX 3090 card.
30
+
31
  ### Training hyperparameters
32
 
33
  The following hyperparameters were used during training:
 
43
  - mixed_precision_training: Native AMP
44
 
45
  ### Training results
46
+ After final epoch:
47
+ {'loss': 0.0472, 'learning_rate': 1.4893617021276598e-06, 'epoch': 4.95}
48
+
49
+ After full completion
50
+ {'train_runtime': 563.2707,
51
+ 'train_samples_per_second': 1.687,
52
+ 'train_steps_per_second': 0.417,
53
+ 'train_loss': 0.8475136074614018,
54
+ 'epoch': 4.95}
55
 
56
 
57
  ### Framework versions