Add more detail on minimum GPU
Browse files
README.md
CHANGED
@@ -135,11 +135,11 @@ See sample configs in [configs](configs) folder or [examples](examples) for quic
|
|
135 |
|
136 |
- loading
|
137 |
```yaml
|
138 |
-
load_4bit: true
|
139 |
load_in_8bit: true
|
140 |
-
|
|
|
141 |
fp16: true
|
142 |
-
tf32: true
|
143 |
```
|
144 |
Note: Repo does not do 4-bit quantization.
|
145 |
|
@@ -183,11 +183,11 @@ gptq_model_v1: false # v1 or v2
|
|
183 |
load_in_8bit: true
|
184 |
|
185 |
# Use CUDA bf16
|
186 |
-
bf16: true # bool or 'full' for `bf16_full_eval
|
187 |
# Use CUDA fp16
|
188 |
fp16: true
|
189 |
# Use CUDA tf32
|
190 |
-
tf32: true
|
191 |
|
192 |
# a list of one or more datasets to finetune the model with
|
193 |
datasets:
|
@@ -286,7 +286,7 @@ weight_decay:
|
|
286 |
# whether to use xformers attention patch https://github.com/facebookresearch/xformers:
|
287 |
xformers_attention:
|
288 |
# whether to use flash attention patch https://github.com/HazyResearch/flash-attention:
|
289 |
-
flash_attention:
|
290 |
|
291 |
# resume from a specific checkpoint dir
|
292 |
resume_from_checkpoint:
|
|
|
135 |
|
136 |
- loading
|
137 |
```yaml
|
|
|
138 |
load_in_8bit: true
|
139 |
+
load_in_8bit: true
|
140 |
+
bf16: true # require >=ampere
|
141 |
fp16: true
|
142 |
+
tf32: true # require >=ampere
|
143 |
```
|
144 |
Note: Repo does not do 4-bit quantization.
|
145 |
|
|
|
183 |
load_in_8bit: true
|
184 |
|
185 |
# Use CUDA bf16
|
186 |
+
bf16: true # bool or 'full' for `bf16_full_eval`. require >=ampere
|
187 |
# Use CUDA fp16
|
188 |
fp16: true
|
189 |
# Use CUDA tf32
|
190 |
+
tf32: true # require >=ampere
|
191 |
|
192 |
# a list of one or more datasets to finetune the model with
|
193 |
datasets:
|
|
|
286 |
# whether to use xformers attention patch https://github.com/facebookresearch/xformers:
|
287 |
xformers_attention:
|
288 |
# whether to use flash attention patch https://github.com/HazyResearch/flash-attention:
|
289 |
+
flash_attention: # require a100 for llama
|
290 |
|
291 |
# resume from a specific checkpoint dir
|
292 |
resume_from_checkpoint:
|