JuanjoLopez19 commited on
Commit
2b69058
1 Parent(s): 4211dda

Training with 90/10 Spanish dataset, 5 epochs, 2 Batch Size, reduce_lr_on_plateau

Browse files
README.md CHANGED
@@ -16,7 +16,7 @@ should probably proofread and complete it, then remove this comment. -->
16
 
17
  This model is a fine-tuned version of [meta-llama/Llama-2-7b-chat-hf](https://huggingface.co/meta-llama/Llama-2-7b-chat-hf) on an unknown dataset.
18
  It achieves the following results on the evaluation set:
19
- - Loss: 1.2459
20
 
21
  ## Model description
22
 
@@ -48,11 +48,11 @@ The following hyperparameters were used during training:
48
 
49
  | Training Loss | Epoch | Step | Validation Loss |
50
  |:-------------:|:------:|:----:|:---------------:|
51
- | 1.3221 | 0.9995 | 1847 | 0.9229 |
52
- | 0.7158 | 1.9989 | 3694 | 0.8894 |
53
- | 0.363 | 2.9984 | 5541 | 0.9671 |
54
- | 0.2104 | 3.9978 | 7388 | 1.1174 |
55
- | 0.39 | 4.9973 | 9235 | 1.2459 |
56
 
57
 
58
  ### Framework versions
 
16
 
17
  This model is a fine-tuned version of [meta-llama/Llama-2-7b-chat-hf](https://huggingface.co/meta-llama/Llama-2-7b-chat-hf) on an unknown dataset.
18
  It achieves the following results on the evaluation set:
19
+ - Loss: 1.2754
20
 
21
  ## Model description
22
 
 
48
 
49
  | Training Loss | Epoch | Step | Validation Loss |
50
  |:-------------:|:------:|:----:|:---------------:|
51
+ | 1.3053 | 0.9995 | 1847 | 0.9206 |
52
+ | 0.7158 | 1.9989 | 3694 | 0.8873 |
53
+ | 0.3506 | 2.9984 | 5541 | 0.9619 |
54
+ | 0.2142 | 3.9978 | 7388 | 1.1203 |
55
+ | 0.3116 | 4.9973 | 9235 | 1.2754 |
56
 
57
 
58
  ### Framework versions
adapter_config.json CHANGED
@@ -23,13 +23,13 @@
23
  "rank_pattern": {},
24
  "revision": null,
25
  "target_modules": [
26
- "o_proj",
27
- "v_proj",
28
- "gate_proj",
29
- "q_proj",
30
  "up_proj",
 
31
  "down_proj",
32
- "k_proj"
 
 
 
33
  ],
34
  "task_type": "CAUSAL_LM",
35
  "use_dora": false,
 
23
  "rank_pattern": {},
24
  "revision": null,
25
  "target_modules": [
 
 
 
 
26
  "up_proj",
27
+ "k_proj",
28
  "down_proj",
29
+ "q_proj",
30
+ "v_proj",
31
+ "gate_proj",
32
+ "o_proj"
33
  ],
34
  "task_type": "CAUSAL_LM",
35
  "use_dora": false,
adapter_model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:3bd9be91216a220040f93b091c1d00d5d815d67496e9d0a97e5bec1a58da3a12
3
  size 1688269144
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:d97a1fd723fc292945f8129a4694d4a02ca7df5702811260ed2dcd90769c7652
3
  size 1688269144
training_args.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:011ec18ec40eb9286a5ab0123e708847240c22c3856cd034d29a50a36580fbeb
3
  size 5048
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:a251263e185cd54161c30d98501e6ab7594498a4add3666d009a3c74f91fa5c9
3
  size 5048