dcbv
/

PEFT
rondlite commited on
Commit
21cab6c
·
1 Parent(s): 00ea6fa

Upload 3 files

Browse files
adapter_config.json CHANGED
@@ -4,7 +4,7 @@
4
  "fan_in_fan_out": false,
5
  "inference_mode": true,
6
  "init_lora_weights": true,
7
- "lora_alpha": 1028,
8
  "lora_dropout": 0.05,
9
  "modules_to_save": null,
10
  "peft_type": "LORA",
 
4
  "fan_in_fan_out": false,
5
  "inference_mode": true,
6
  "init_lora_weights": true,
7
+ "lora_alpha": 1024,
8
  "lora_dropout": 0.05,
9
  "modules_to_save": null,
10
  "peft_type": "LORA",
adapter_model.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:18b41b3f6ad2bff148d99a2206c09a1d08ec0df76984dfcb9f1d17d59efc1c5c
3
- size 154140672
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:e9e22a25d5179569fcae8c2d59d646d5d464461fb8b83e0dd224dcb4f2bb5686
3
+ size 112721920
training_parameters.json CHANGED
@@ -1 +1 @@
1
- {"lora_name": "charluv-lora", "always_override": true, "save_steps": 0.0, "micro_batch_size": 4, "batch_size": 128, "epochs": 3.0, "learning_rate": "3e-5", "lr_scheduler_type": "cosine_with_restarts", "lora_rank": 512, "lora_alpha": 1028, "lora_dropout": 0.05, "cutoff_len": 256, "dataset": "None", "eval_dataset": "None", "format": "None", "eval_steps": 100.0, "raw_text_file": "aivo", "overlap_len": 128, "newline_favor_len": 128, "do_shuffle": false, "higher_rank_limit": false, "warmup_steps": 100.0, "optimizer": "adamw_torch"}
 
1
+ {"lora_name": "charluv-lora", "always_override": true, "save_steps": 0.0, "micro_batch_size": 4, "batch_size": 128, "epochs": 8.0, "learning_rate": "3e-5", "lr_scheduler_type": "cosine_with_restarts", "lora_rank": 512, "lora_alpha": 1024, "lora_dropout": 0.05, "cutoff_len": 256, "dataset": "None", "eval_dataset": "None", "format": "None", "eval_steps": 100.0, "raw_text_file": "aivo", "overlap_len": 128, "newline_favor_len": 128, "do_shuffle": false, "higher_rank_limit": false, "warmup_steps": 100.0, "optimizer": "adamw_torch"}