Nanobit commited on
Commit
e65c203
1 Parent(s): 1377400

Add more detail on minimum GPU

Browse files
Files changed (1) hide show
  1. README.md +6 -6
README.md CHANGED
@@ -135,11 +135,11 @@ See sample configs in [configs](configs) folder or [examples](examples) for quic
135
 
136
  - loading
137
  ```yaml
138
- load_4bit: true
139
  load_in_8bit: true
140
- bf16: true
 
141
  fp16: true
142
- tf32: true
143
  ```
144
  Note: Repo does not do 4-bit quantization.
145
 
@@ -183,11 +183,11 @@ gptq_model_v1: false # v1 or v2
183
  load_in_8bit: true
184
 
185
  # Use CUDA bf16
186
- bf16: true # bool or 'full' for `bf16_full_eval`
187
  # Use CUDA fp16
188
  fp16: true
189
  # Use CUDA tf32
190
- tf32: true
191
 
192
  # a list of one or more datasets to finetune the model with
193
  datasets:
@@ -286,7 +286,7 @@ weight_decay:
286
  # whether to use xformers attention patch https://github.com/facebookresearch/xformers:
287
  xformers_attention:
288
  # whether to use flash attention patch https://github.com/HazyResearch/flash-attention:
289
- flash_attention:
290
 
291
  # resume from a specific checkpoint dir
292
  resume_from_checkpoint:
 
135
 
136
  - loading
137
  ```yaml
 
138
  load_in_8bit: true
139
+ load_in_8bit: true
140
+ bf16: true # require >=ampere
141
  fp16: true
142
+ tf32: true # require >=ampere
143
  ```
144
  Note: Repo does not do 4-bit quantization.
145
 
 
183
  load_in_8bit: true
184
 
185
  # Use CUDA bf16
186
+ bf16: true # bool or 'full' for `bf16_full_eval`. require >=ampere
187
  # Use CUDA fp16
188
  fp16: true
189
  # Use CUDA tf32
190
+ tf32: true # require >=ampere
191
 
192
  # a list of one or more datasets to finetune the model with
193
  datasets:
 
286
  # whether to use xformers attention patch https://github.com/facebookresearch/xformers:
287
  xformers_attention:
288
  # whether to use flash attention patch https://github.com/HazyResearch/flash-attention:
289
+ flash_attention: # require a100 for llama
290
 
291
  # resume from a specific checkpoint dir
292
  resume_from_checkpoint: