Commit
·
de52ddd
1
Parent(s):
12103d6
Update README.md
Browse files
README.md
CHANGED
@@ -3,31 +3,3 @@ license: mit
|
|
3 |
datasets:
|
4 |
- yahma/alpaca-cleaned
|
5 |
---
|
6 |
-
|
7 |
-
This repo contains a low-rank adapter for LLaMA-7b
|
8 |
-
fit on the [Stanford Alpaca](https://github.com/tatsu-lab/stanford_alpaca) dataset.
|
9 |
-
|
10 |
-
This version of the weights was trained with the following hyperparameters:
|
11 |
-
|
12 |
-
- Epochs: 10 (load from best epoch)
|
13 |
-
- Batch size: 128
|
14 |
-
- Cutoff length: 512
|
15 |
-
- Learning rate: 3e-4
|
16 |
-
- Lora _r_: 16
|
17 |
-
- Lora target modules: q_proj, k_proj, v_proj, o_proj
|
18 |
-
|
19 |
-
That is:
|
20 |
-
|
21 |
-
```
|
22 |
-
python finetune.py \
|
23 |
-
--base_model='decapoda-research/llama-7b-hf' \
|
24 |
-
--num_epochs=10 \
|
25 |
-
--cutoff_len=512 \
|
26 |
-
--group_by_length \
|
27 |
-
--output_dir='./lora-alpaca-512-qkvo' \
|
28 |
-
--lora_target_modules='[q_proj,k_proj,v_proj,o_proj]' \
|
29 |
-
--lora_r=16 \
|
30 |
-
--micro_batch_size=8
|
31 |
-
```
|
32 |
-
|
33 |
-
Instructions for running it can be found at https://github.com/tloen/alpaca-lora.
|
|
|
3 |
datasets:
|
4 |
- yahma/alpaca-cleaned
|
5 |
---
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|