alendra1945 commited on
Commit
de52ddd
·
1 Parent(s): 12103d6

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +0 -28
README.md CHANGED
@@ -3,31 +3,3 @@ license: mit
3
  datasets:
4
  - yahma/alpaca-cleaned
5
  ---
6
-
7
- This repo contains a low-rank adapter for LLaMA-7b
8
- fit on the [Stanford Alpaca](https://github.com/tatsu-lab/stanford_alpaca) dataset.
9
-
10
- This version of the weights was trained with the following hyperparameters:
11
-
12
- - Epochs: 10 (load from best epoch)
13
- - Batch size: 128
14
- - Cutoff length: 512
15
- - Learning rate: 3e-4
16
- - Lora _r_: 16
17
- - Lora target modules: q_proj, k_proj, v_proj, o_proj
18
-
19
- That is:
20
-
21
- ```
22
- python finetune.py \
23
- --base_model='decapoda-research/llama-7b-hf' \
24
- --num_epochs=10 \
25
- --cutoff_len=512 \
26
- --group_by_length \
27
- --output_dir='./lora-alpaca-512-qkvo' \
28
- --lora_target_modules='[q_proj,k_proj,v_proj,o_proj]' \
29
- --lora_r=16 \
30
- --micro_batch_size=8
31
- ```
32
-
33
- Instructions for running it can be found at https://github.com/tloen/alpaca-lora.
 
3
  datasets:
4
  - yahma/alpaca-cleaned
5
  ---