terry69 commited on
Commit
1a8a5b5
1 Parent(s): b30bca5

Model save

Browse files
README.md CHANGED
@@ -1,16 +1,11 @@
1
  ---
2
  license: apache-2.0
3
- base_model: mistralai/Mistral-7B-Instruct-v0.2
4
  tags:
5
- - alignment-handbook
6
  - trl
7
  - sft
8
  - generated_from_trainer
9
- - trl
10
- - sft
11
- - generated_from_trainer
12
- datasets:
13
- - preference-data
14
  model-index:
15
  - name: preference_p0.1_seed42_level2_raremixbatch16
16
  results: []
@@ -21,9 +16,9 @@ should probably proofread and complete it, then remove this comment. -->
21
 
22
  # preference_p0.1_seed42_level2_raremixbatch16
23
 
24
- This model is a fine-tuned version of [mistralai/Mistral-7B-Instruct-v0.2](https://huggingface.co/mistralai/Mistral-7B-Instruct-v0.2) on the preference-data dataset.
25
  It achieves the following results on the evaluation set:
26
- - Loss: 0.2881
27
 
28
  ## Model description
29
 
@@ -59,12 +54,13 @@ The following hyperparameters were used during training:
59
 
60
  | Training Loss | Epoch | Step | Validation Loss |
61
  |:-------------:|:-----:|:----:|:---------------:|
62
- | 0.3161 | 1.0 | 2414 | 0.2881 |
63
 
64
 
65
  ### Framework versions
66
 
67
- - Transformers 4.43.4
 
68
  - Pytorch 2.3.1+cu121
69
  - Datasets 2.19.1
70
- - Tokenizers 0.19.1
 
1
  ---
2
  license: apache-2.0
3
+ library_name: peft
4
  tags:
 
5
  - trl
6
  - sft
7
  - generated_from_trainer
8
+ base_model: mistralai/Mistral-7B-Instruct-v0.2
 
 
 
 
9
  model-index:
10
  - name: preference_p0.1_seed42_level2_raremixbatch16
11
  results: []
 
16
 
17
  # preference_p0.1_seed42_level2_raremixbatch16
18
 
19
+ This model is a fine-tuned version of [mistralai/Mistral-7B-Instruct-v0.2](https://huggingface.co/mistralai/Mistral-7B-Instruct-v0.2) on the None dataset.
20
  It achieves the following results on the evaluation set:
21
+ - Loss: 0.8599
22
 
23
  ## Model description
24
 
 
54
 
55
  | Training Loss | Epoch | Step | Validation Loss |
56
  |:-------------:|:-----:|:----:|:---------------:|
57
+ | 0.8409 | 1.0 | 6181 | 0.8599 |
58
 
59
 
60
  ### Framework versions
61
 
62
+ - PEFT 0.11.1
63
+ - Transformers 4.44.2
64
  - Pytorch 2.3.1+cu121
65
  - Datasets 2.19.1
66
+ - Tokenizers 0.19.1
adapter_config.json ADDED
@@ -0,0 +1,35 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "alpha_pattern": {},
3
+ "auto_mapping": null,
4
+ "base_model_name_or_path": "mistralai/Mistral-7B-Instruct-v0.2",
5
+ "bias": "none",
6
+ "fan_in_fan_out": false,
7
+ "inference_mode": true,
8
+ "init_lora_weights": true,
9
+ "layer_replication": null,
10
+ "layers_pattern": null,
11
+ "layers_to_transform": null,
12
+ "loftq_config": {},
13
+ "lora_alpha": 64,
14
+ "lora_dropout": 0.1,
15
+ "megatron_config": null,
16
+ "megatron_core": "megatron.core",
17
+ "modules_to_save": null,
18
+ "peft_type": "LORA",
19
+ "r": 32,
20
+ "rank_pattern": {},
21
+ "revision": null,
22
+ "target_modules": [
23
+ "k_proj",
24
+ "w2",
25
+ "w1",
26
+ "v_proj",
27
+ "q_proj",
28
+ "w3",
29
+ "lm_head",
30
+ "o_proj"
31
+ ],
32
+ "task_type": "CAUSAL_LM",
33
+ "use_dora": false,
34
+ "use_rslora": false
35
+ }
adapter_model.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:63b28d7e6592845b2e369912a7b0ab05112233cb2adcff154784d02a2c0d9318
3
+ size 319015728
all_results.json CHANGED
@@ -1,14 +1,9 @@
1
  {
2
  "epoch": 1.0,
3
- "eval_loss": 0.28806865215301514,
4
- "eval_runtime": 1.3149,
5
- "eval_samples": 10,
6
- "eval_samples_per_second": 2.282,
7
- "eval_steps_per_second": 0.761,
8
- "total_flos": 252721244405760.0,
9
- "train_loss": 0.5156875643339054,
10
- "train_runtime": 24515.4963,
11
  "train_samples": 98881,
12
- "train_samples_per_second": 1.575,
13
- "train_steps_per_second": 0.098
14
  }
 
1
  {
2
  "epoch": 1.0,
3
+ "total_flos": 2571089024581632.0,
4
+ "train_loss": 0.8638364601301117,
5
+ "train_runtime": 28571.7374,
 
 
 
 
 
6
  "train_samples": 98881,
7
+ "train_samples_per_second": 3.461,
8
+ "train_steps_per_second": 0.216
9
  }
runs/Sep14_01-24-34_COE-CS-sv004/events.out.tfevents.1726277286.COE-CS-sv004.1678647.0 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:d7476b0540fc0ba01e51c18044cd975822ac8088765b3366bc8120a38bca7c6a
3
+ size 266794
tokenizer_config.json CHANGED
@@ -30,11 +30,11 @@
30
  },
31
  "additional_special_tokens": [],
32
  "bos_token": "<s>",
33
- "chat_template": "{% for message in messages %}\n{% if message['role'] == 'user' %}\n{{ '<|user|>\n' + message['content'] + eos_token }}\n{% elif message['role'] == 'system' %}\n{{ '<|system|>\n' + message['content'] + eos_token }}\n{% elif message['role'] == 'assistant' %}\n{{ '<|assistant|>\n' + message['content'] + eos_token }}\n{% endif %}\n{% if loop.last and add_generation_prompt %}\n{{ '<|assistant|>' }}\n{% endif %}\n{% endfor %}",
34
  "clean_up_tokenization_spaces": false,
35
  "eos_token": "</s>",
36
  "legacy": false,
37
- "model_max_length": 2048,
38
  "pad_token": "</s>",
39
  "sp_model_kwargs": {},
40
  "spaces_between_special_tokens": false,
 
30
  },
31
  "additional_special_tokens": [],
32
  "bos_token": "<s>",
33
+ "chat_template": "{%- if messages[0]['role'] == 'system' %}\n {%- set system_message = messages[0]['content'] %}\n {%- set loop_messages = messages[1:] %}\n{%- else %}\n {%- set loop_messages = messages %}\n{%- endif %}\n\n{{- bos_token }}\n{%- for message in loop_messages %}\n {%- if (message['role'] == 'user') != (loop.index0 % 2 == 0) %}\n {{- raise_exception('After the optional system message, conversation roles must alternate user/assistant/user/assistant/...') }}\n {%- endif %}\n {%- if message['role'] == 'user' %}\n {%- if loop.first and system_message is defined %}\n {{- ' [INST] ' + system_message + '\\n\\n' + message['content'] + ' [/INST]' }}\n {%- else %}\n {{- ' [INST] ' + message['content'] + ' [/INST]' }}\n {%- endif %}\n {%- elif message['role'] == 'assistant' %}\n {{- ' ' + message['content'] + eos_token}}\n {%- else %}\n {{- raise_exception('Only user and assistant roles are supported, with the exception of an initial optional system message!') }}\n {%- endif %}\n{%- endfor %}\n",
34
  "clean_up_tokenization_spaces": false,
35
  "eos_token": "</s>",
36
  "legacy": false,
37
+ "model_max_length": 1000000000000000019884624838656,
38
  "pad_token": "</s>",
39
  "sp_model_kwargs": {},
40
  "spaces_between_special_tokens": false,
train_results.json CHANGED
@@ -1,9 +1,9 @@
1
  {
2
  "epoch": 1.0,
3
- "total_flos": 252721244405760.0,
4
- "train_loss": 0.5156875643339054,
5
- "train_runtime": 24515.4963,
6
  "train_samples": 98881,
7
- "train_samples_per_second": 1.575,
8
- "train_steps_per_second": 0.098
9
  }
 
1
  {
2
  "epoch": 1.0,
3
+ "total_flos": 2571089024581632.0,
4
+ "train_loss": 0.8638364601301117,
5
+ "train_runtime": 28571.7374,
6
  "train_samples": 98881,
7
+ "train_samples_per_second": 3.461,
8
+ "train_steps_per_second": 0.216
9
  }
trainer_state.json CHANGED
The diff for this file is too large to render. See raw diff
 
training_args.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:b187636bdae2f5fd2921d35e55c44fba16fb41f45d96bba8ece972c877b93c3a
3
- size 6584
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:2d2cfab61ea7d4ddc928c8af1ae1777847f06be2085febf03b4016a7ff48a13a
3
+ size 7032