Nanobit commited on
Commit
ba9ac72
β€’
1 Parent(s): db73b94

Update quickstart. Add common error and contribution section.

Browse files
Files changed (1) hide show
  1. README.md +24 -5
README.md CHANGED
@@ -22,7 +22,7 @@
22
  | mpt | βœ… | ❌ | ❌ | ❌ | ❌ | ❓ |
23
 
24
 
25
- ## Quick start
26
 
27
  **Requirements**: Python 3.9.
28
 
@@ -32,12 +32,15 @@ git clone https://github.com/OpenAccess-AI-Collective/axolotl
32
  pip3 install -e .[int4]
33
 
34
  accelerate config
35
- accelerate launch scripts/finetune.py examples/4bit-lora-7b/config.yml
36
- ```
37
 
 
 
38
 
 
 
 
39
 
40
- ## Requirements and Installation
41
 
42
  ### Environment
43
 
@@ -108,6 +111,8 @@ Have dataset(s) in one of the following format (JSONL recommended):
108
  {"text": "..."}
109
  ```
110
 
 
 
111
  </details>
112
 
113
  Optionally, download some datasets, see [data/README.md](data/README.md)
@@ -309,6 +314,7 @@ Configure accelerate
309
  ```bash
310
  accelerate config
311
 
 
312
  # nano ~/.cache/huggingface/accelerate/default_config.yaml
313
  ```
314
 
@@ -330,5 +336,18 @@ If you are inferencing a pretrained LORA, pass
330
 
331
  ### Merge LORA to base (Dev branch πŸ”§ )
332
 
333
- Add `--merge_lora --lora_model_dir="path/to/lora"` flag to train command above
 
 
 
 
 
 
 
 
 
 
 
 
334
 
 
 
22
  | mpt | βœ… | ❌ | ❌ | ❌ | ❌ | ❓ |
23
 
24
 
25
+ ## Quickstart ⚑
26
 
27
  **Requirements**: Python 3.9.
28
 
 
32
  pip3 install -e .[int4]
33
 
34
  accelerate config
 
 
35
 
36
+ # finetune
37
+ accelerate launch scripts/finetune.py examples/4bit-lora-7b/config.yml
38
 
39
+ # inference
40
+ accelerate launch scripts/finetune.py examples/4bit-lora-7b/config.yml --inference
41
+ ```
42
 
43
+ ## Installation
44
 
45
  ### Environment
46
 
 
111
  {"text": "..."}
112
  ```
113
 
114
+ > Have some new format to propose? Check if it's already defined in [data.py](src/axolotl/utils/data.py) in `dev` branch!
115
+
116
  </details>
117
 
118
  Optionally, download some datasets, see [data/README.md](data/README.md)
 
314
  ```bash
315
  accelerate config
316
 
317
+ # Edit manually
318
  # nano ~/.cache/huggingface/accelerate/default_config.yaml
319
  ```
320
 
 
336
 
337
  ### Merge LORA to base (Dev branch πŸ”§ )
338
 
339
+ Add below flag to train command above
340
+
341
+ ```bash
342
+ --merge_lora --lora_model_dir="./completed-model"
343
+ ```
344
+
345
+ ## Common Errors 🧰
346
+
347
+ - Cuda out of memory: Please reduce `micro_batch_size` and/or `eval_batch_size`
348
+
349
+ ## Contributing 🀝
350
+
351
+ Bugs? Please check for open issue else create a new [Issue](https://github.com/OpenAccess-AI-Collective/axolotl/issues/new).
352
 
353
+ PRs are **greatly welcome**!