Upload 3 files
Browse files- adapter_config.json +1 -1
- adapter_model.bin +2 -2
- training_parameters.json +1 -1
adapter_config.json
CHANGED
@@ -4,7 +4,7 @@
|
|
4 |
"fan_in_fan_out": false,
|
5 |
"inference_mode": true,
|
6 |
"init_lora_weights": true,
|
7 |
-
"lora_alpha":
|
8 |
"lora_dropout": 0.05,
|
9 |
"modules_to_save": null,
|
10 |
"peft_type": "LORA",
|
|
|
4 |
"fan_in_fan_out": false,
|
5 |
"inference_mode": true,
|
6 |
"init_lora_weights": true,
|
7 |
+
"lora_alpha": 1024,
|
8 |
"lora_dropout": 0.05,
|
9 |
"modules_to_save": null,
|
10 |
"peft_type": "LORA",
|
adapter_model.bin
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
-
size
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:e9e22a25d5179569fcae8c2d59d646d5d464461fb8b83e0dd224dcb4f2bb5686
|
3 |
+
size 112721920
|
training_parameters.json
CHANGED
@@ -1 +1 @@
|
|
1 |
-
{"lora_name": "charluv-lora", "always_override": true, "save_steps": 0.0, "micro_batch_size": 4, "batch_size": 128, "epochs":
|
|
|
1 |
+
{"lora_name": "charluv-lora", "always_override": true, "save_steps": 0.0, "micro_batch_size": 4, "batch_size": 128, "epochs": 8.0, "learning_rate": "3e-5", "lr_scheduler_type": "cosine_with_restarts", "lora_rank": 512, "lora_alpha": 1024, "lora_dropout": 0.05, "cutoff_len": 256, "dataset": "None", "eval_dataset": "None", "format": "None", "eval_steps": 100.0, "raw_text_file": "aivo", "overlap_len": 128, "newline_favor_len": 128, "do_shuffle": false, "higher_rank_limit": false, "warmup_steps": 100.0, "optimizer": "adamw_torch"}
|