Update README.md
Browse files
README.md
CHANGED
@@ -18,7 +18,7 @@ This version of the weights was trained with the following hyperparameters:
|
|
18 |
- Batch size: 128
|
19 |
- Micro batch size: 4
|
20 |
- Learning rate: 3e-4
|
21 |
-
- Lora _r_:
|
22 |
- Lora target modules: query_key_value
|
23 |
|
24 |
You can reproduce using this repository:
|
@@ -29,10 +29,9 @@ Make sure you install requirements and finetune using this command using the fol
|
|
29 |
|
30 |
```
|
31 |
python finetune.py \
|
32 |
-
|
33 |
-
|
34 |
-
|
35 |
-
|
36 |
-
|
37 |
-
--micro-batch-size=4
|
38 |
```
|
|
|
18 |
- Batch size: 128
|
19 |
- Micro batch size: 4
|
20 |
- Learning rate: 3e-4
|
21 |
+
- Lora _r_: 8
|
22 |
- Lora target modules: query_key_value
|
23 |
|
24 |
You can reproduce using this repository:
|
|
|
29 |
|
30 |
```
|
31 |
python finetune.py \
|
32 |
+
--base-model tiiuae/falcon-40b --lora-target-modules query_key_value \
|
33 |
+
--data-path sahil2801/CodeAlpaca-20k --output-dir ./lora-alpaca-code \
|
34 |
+
--batch-size 128 --micro-batch-size 4 --eval-limit 45 \
|
35 |
+
--eval-file code_eval.jsonl --wandb-project jerboa --wandb-log-model \
|
36 |
+
--wandb-watch gradients --num-epochs 2
|
|
|
37 |
```
|