Astris commited on
Commit
175dc21
1 Parent(s): 915f6e4

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +6 -2
README.md CHANGED
@@ -1,8 +1,6 @@
1
  ---
2
  language:
3
  - en
4
- tags:
5
- - not-for-all-audiences
6
  ---
7
  # Summary
8
  The name is self-explanatory. This LoRA was trained on 50MB of text taken from Archive Of Our Own (AO3). In total, 1441 stories were selected from the Furry fandom category. I don't remember what filters I used.
@@ -17,11 +15,17 @@ The name is self-explanatory. This LoRA was trained on 50MB of text taken from A
17
  - Targeted modules: Q, K, V, O, Gate, Up, Down
18
  - NEFTune alpha: 10 (to try to reduce overfitting)
19
  - Learning rate: 1e-4
 
20
 
21
  # Model Settings
22
  - Base model: Mistral 7B
23
  - Data Type: BF16, 4 bit quantization (thanks BitsandBytes)
24
 
 
 
 
 
 
25
  # Software and Hardware
26
  - Unsloth was used to speed up training.
27
  - Training was done on 1x RTX 3090 (with 24 GB of VRAM) and took 11 hours.
 
1
  ---
2
  language:
3
  - en
 
 
4
  ---
5
  # Summary
6
  The name is self-explanatory. This LoRA was trained on 50MB of text taken from Archive Of Our Own (AO3). In total, 1441 stories were selected from the Furry fandom category. I don't remember what filters I used.
 
15
  - Targeted modules: Q, K, V, O, Gate, Up, Down
16
  - NEFTune alpha: 10 (to try to reduce overfitting)
17
  - Learning rate: 1e-4
18
+ - Dropout: 0 (unsloth doesn't support LoRA dropout)
19
 
20
  # Model Settings
21
  - Base model: Mistral 7B
22
  - Data Type: BF16, 4 bit quantization (thanks BitsandBytes)
23
 
24
+ # Misc Settings
25
+ - Batch size: 2
26
+ - Gradient Accumulation steps: 16
27
+ - LR Scheduler: Linear
28
+
29
  # Software and Hardware
30
  - Unsloth was used to speed up training.
31
  - Training was done on 1x RTX 3090 (with 24 GB of VRAM) and took 11 hours.