nblinh commited on
Commit
dff5775
1 Parent(s): 3efb08d

End of training

Browse files
README.md CHANGED
@@ -78,7 +78,7 @@ resume_from_checkpoint: null
78
  s2_attention: null
79
  sample_packing: false
80
  saves_per_epoch: 1
81
- sequence_len: 1024
82
  strict: false
83
  tf32: false
84
  tokenizer_type: AutoTokenizer
@@ -103,7 +103,7 @@ xformers_attention: true
103
 
104
  This model is a fine-tuned version of [migtissera/Tess-v2.5-Phi-3-medium-128k-14B](https://huggingface.co/migtissera/Tess-v2.5-Phi-3-medium-128k-14B) on the None dataset.
105
  It achieves the following results on the evaluation set:
106
- - Loss: 2.9300
107
 
108
  ## Model description
109
 
@@ -135,7 +135,7 @@ The following hyperparameters were used during training:
135
 
136
  | Training Loss | Epoch | Step | Validation Loss |
137
  |:-------------:|:------:|:----:|:---------------:|
138
- | 3.2576 | 0.0002 | 10 | 2.9300 |
139
 
140
 
141
  ### Framework versions
 
78
  s2_attention: null
79
  sample_packing: false
80
  saves_per_epoch: 1
81
+ sequence_len: 2048
82
  strict: false
83
  tf32: false
84
  tokenizer_type: AutoTokenizer
 
103
 
104
  This model is a fine-tuned version of [migtissera/Tess-v2.5-Phi-3-medium-128k-14B](https://huggingface.co/migtissera/Tess-v2.5-Phi-3-medium-128k-14B) on the None dataset.
105
  It achieves the following results on the evaluation set:
106
+ - Loss: 2.9233
107
 
108
  ## Model description
109
 
 
135
 
136
  | Training Loss | Epoch | Step | Validation Loss |
137
  |:-------------:|:------:|:----:|:---------------:|
138
+ | 3.2514 | 0.0002 | 10 | 2.9233 |
139
 
140
 
141
  ### Framework versions
adapter_config.json CHANGED
@@ -20,10 +20,10 @@
20
  "rank_pattern": {},
21
  "revision": null,
22
  "target_modules": [
23
- "down_proj",
24
- "gate_up_proj",
25
  "qkv_proj",
26
- "o_proj"
 
 
27
  ],
28
  "task_type": "CAUSAL_LM",
29
  "use_dora": false,
 
20
  "rank_pattern": {},
21
  "revision": null,
22
  "target_modules": [
 
 
23
  "qkv_proj",
24
+ "o_proj",
25
+ "down_proj",
26
+ "gate_up_proj"
27
  ],
28
  "task_type": "CAUSAL_LM",
29
  "use_dora": false,
adapter_model.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:3ebf79661b002aedc25155520a0d6aeb227e31cab4e96e43f38c2841fc2daf26
3
  size 222938570
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:6c5e7d03ace063deabbde439ed04e33123f49f1c6b1e7e16c2b022d2c4997c33
3
  size 222938570
adapter_model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:353ce181c025137d557478b504ca276623dddebd142c6b81c068c6e199f36731
3
  size 222865880
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:e8e22e420389d37fcdc287724c242e711b1581e4c8908f067573e329bb442773
3
  size 222865880
training_args.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:b78a686c48eeb62bd6518cf77d58567ef22b54d9ff01c99be6e2ee1e84b0faca
3
  size 6776
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:0148c0006a77c50cfa094a33b984c86682fef78d700a4ae2fce51d620b338d5a
3
  size 6776