terry69 commited on
Commit
7a3caa3
1 Parent(s): 74b349a

Model save

Browse files
README.md CHANGED
@@ -1,14 +1,12 @@
1
  ---
 
2
  license: apache-2.0
3
- library_name: peft
4
  tags:
5
- - alignment-handbook
6
  - trl
7
  - sft
 
8
  - generated_from_trainer
9
- base_model: mistralai/Mistral-7B-Instruct-v0.2
10
- datasets:
11
- - preference-data
12
  model-index:
13
  - name: preference_p0.1_seed42_level2_raremixbatch16
14
  results: []
@@ -19,9 +17,9 @@ should probably proofread and complete it, then remove this comment. -->
19
 
20
  # preference_p0.1_seed42_level2_raremixbatch16
21
 
22
- This model is a fine-tuned version of [mistralai/Mistral-7B-Instruct-v0.2](https://huggingface.co/mistralai/Mistral-7B-Instruct-v0.2) on the preference-data dataset.
23
  It achieves the following results on the evaluation set:
24
- - Loss: 0.8599
25
 
26
  ## Model description
27
 
@@ -57,13 +55,12 @@ The following hyperparameters were used during training:
57
 
58
  | Training Loss | Epoch | Step | Validation Loss |
59
  |:-------------:|:-----:|:----:|:---------------:|
60
- | 0.8409 | 1.0 | 6181 | 0.8599 |
61
 
62
 
63
  ### Framework versions
64
 
65
- - PEFT 0.11.1
66
  - Transformers 4.44.2
67
  - Pytorch 2.3.1+cu121
68
  - Datasets 2.19.1
69
- - Tokenizers 0.19.1
 
1
  ---
2
+ library_name: transformers
3
  license: apache-2.0
4
+ base_model: mistralai/Mistral-7B-Instruct-v0.2
5
  tags:
 
6
  - trl
7
  - sft
8
+ - alignment-handbook
9
  - generated_from_trainer
 
 
 
10
  model-index:
11
  - name: preference_p0.1_seed42_level2_raremixbatch16
12
  results: []
 
17
 
18
  # preference_p0.1_seed42_level2_raremixbatch16
19
 
20
+ This model is a fine-tuned version of [mistralai/Mistral-7B-Instruct-v0.2](https://huggingface.co/mistralai/Mistral-7B-Instruct-v0.2) on the None dataset.
21
  It achieves the following results on the evaluation set:
22
+ - Loss: 0.2881
23
 
24
  ## Model description
25
 
 
55
 
56
  | Training Loss | Epoch | Step | Validation Loss |
57
  |:-------------:|:-----:|:----:|:---------------:|
58
+ | 0.3161 | 1.0 | 2414 | 0.2881 |
59
 
60
 
61
  ### Framework versions
62
 
 
63
  - Transformers 4.44.2
64
  - Pytorch 2.3.1+cu121
65
  - Datasets 2.19.1
66
+ - Tokenizers 0.19.1
all_results.json CHANGED
@@ -1,14 +1,14 @@
1
  {
2
  "epoch": 1.0,
3
- "eval_loss": 0.8599082231521606,
4
- "eval_runtime": 2.8502,
5
  "eval_samples": 10,
6
- "eval_samples_per_second": 3.508,
7
- "eval_steps_per_second": 1.053,
8
- "total_flos": 2571089024581632.0,
9
- "train_loss": 0.8638364601301117,
10
- "train_runtime": 28571.7374,
11
  "train_samples": 98881,
12
- "train_samples_per_second": 3.461,
13
- "train_steps_per_second": 0.216
14
  }
 
1
  {
2
  "epoch": 1.0,
3
+ "eval_loss": 0.28806865215301514,
4
+ "eval_runtime": 1.3149,
5
  "eval_samples": 10,
6
+ "eval_samples_per_second": 2.282,
7
+ "eval_steps_per_second": 0.761,
8
+ "total_flos": 252721244405760.0,
9
+ "train_loss": 0.0,
10
+ "train_runtime": 0.0097,
11
  "train_samples": 98881,
12
+ "train_samples_per_second": 3991047.139,
13
+ "train_steps_per_second": 249446.905
14
  }
config.json CHANGED
@@ -22,6 +22,6 @@
22
  "tie_word_embeddings": false,
23
  "torch_dtype": "bfloat16",
24
  "transformers_version": "4.44.2",
25
- "use_cache": true,
26
  "vocab_size": 32000
27
  }
 
22
  "tie_word_embeddings": false,
23
  "torch_dtype": "bfloat16",
24
  "transformers_version": "4.44.2",
25
+ "use_cache": false,
26
  "vocab_size": 32000
27
  }
eval_results.json CHANGED
@@ -1,8 +1,8 @@
1
  {
2
  "epoch": 1.0,
3
- "eval_loss": 0.8599082231521606,
4
- "eval_runtime": 2.8502,
5
  "eval_samples": 10,
6
- "eval_samples_per_second": 3.508,
7
- "eval_steps_per_second": 1.053
8
  }
 
1
  {
2
  "epoch": 1.0,
3
+ "eval_loss": 0.28806865215301514,
4
+ "eval_runtime": 1.3149,
5
  "eval_samples": 10,
6
+ "eval_samples_per_second": 2.282,
7
+ "eval_steps_per_second": 0.761
8
  }
generation_config.json CHANGED
@@ -2,5 +2,5 @@
2
  "_from_model_config": true,
3
  "bos_token_id": 1,
4
  "eos_token_id": 2,
5
- "transformers_version": "4.43.4"
6
  }
 
2
  "_from_model_config": true,
3
  "bos_token_id": 1,
4
  "eos_token_id": 2,
5
+ "transformers_version": "4.44.2"
6
  }
runs/Sep24_11-17-32_COE-CS-sv003/events.out.tfevents.1727176715.COE-CS-sv003.779460.0 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:071577acc062c85ef97090426c62030ff7e01f6d95f6c06337d56e8cd48a19e1
3
+ size 5620
tokenizer_config.json CHANGED
@@ -30,11 +30,11 @@
30
  },
31
  "additional_special_tokens": [],
32
  "bos_token": "<s>",
33
- "chat_template": "{%- if messages[0]['role'] == 'system' %}\n {%- set system_message = messages[0]['content'] %}\n {%- set loop_messages = messages[1:] %}\n{%- else %}\n {%- set loop_messages = messages %}\n{%- endif %}\n\n{{- bos_token }}\n{%- for message in loop_messages %}\n {%- if (message['role'] == 'user') != (loop.index0 % 2 == 0) %}\n {{- raise_exception('After the optional system message, conversation roles must alternate user/assistant/user/assistant/...') }}\n {%- endif %}\n {%- if message['role'] == 'user' %}\n {%- if loop.first and system_message is defined %}\n {{- ' [INST] ' + system_message + '\\n\\n' + message['content'] + ' [/INST]' }}\n {%- else %}\n {{- ' [INST] ' + message['content'] + ' [/INST]' }}\n {%- endif %}\n {%- elif message['role'] == 'assistant' %}\n {{- ' ' + message['content'] + eos_token}}\n {%- else %}\n {{- raise_exception('Only user and assistant roles are supported, with the exception of an initial optional system message!') }}\n {%- endif %}\n{%- endfor %}\n",
34
  "clean_up_tokenization_spaces": false,
35
  "eos_token": "</s>",
36
  "legacy": false,
37
- "model_max_length": 1000000000000000019884624838656,
38
  "pad_token": "</s>",
39
  "sp_model_kwargs": {},
40
  "spaces_between_special_tokens": false,
 
30
  },
31
  "additional_special_tokens": [],
32
  "bos_token": "<s>",
33
+ "chat_template": "{% for message in messages %}\n{% if message['role'] == 'user' %}\n{{ '<|user|>\n' + message['content'] + eos_token }}\n{% elif message['role'] == 'system' %}\n{{ '<|system|>\n' + message['content'] + eos_token }}\n{% elif message['role'] == 'assistant' %}\n{{ '<|assistant|>\n' + message['content'] + eos_token }}\n{% endif %}\n{% if loop.last and add_generation_prompt %}\n{{ '<|assistant|>' }}\n{% endif %}\n{% endfor %}",
34
  "clean_up_tokenization_spaces": false,
35
  "eos_token": "</s>",
36
  "legacy": false,
37
+ "model_max_length": 2048,
38
  "pad_token": "</s>",
39
  "sp_model_kwargs": {},
40
  "spaces_between_special_tokens": false,
train_results.json CHANGED
@@ -1,9 +1,9 @@
1
  {
2
  "epoch": 1.0,
3
- "total_flos": 2571089024581632.0,
4
- "train_loss": 0.8638364601301117,
5
- "train_runtime": 28571.7374,
6
  "train_samples": 98881,
7
- "train_samples_per_second": 3.461,
8
- "train_steps_per_second": 0.216
9
  }
 
1
  {
2
  "epoch": 1.0,
3
+ "total_flos": 252721244405760.0,
4
+ "train_loss": 0.0,
5
+ "train_runtime": 0.0097,
6
  "train_samples": 98881,
7
+ "train_samples_per_second": 3991047.139,
8
+ "train_steps_per_second": 249446.905
9
  }
trainer_state.json CHANGED
The diff for this file is too large to render. See raw diff
 
training_args.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:2d2cfab61ea7d4ddc928c8af1ae1777847f06be2085febf03b4016a7ff48a13a
3
  size 7032
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:cf2d01fd7c3b20690fa23c87f64bf43148f8884951c9cf6af8410d6fcbb51ec7
3
  size 7032