Aravindan commited on
Commit
8e108a8
1 Parent(s): 8ade26b

End of training

Browse files
README.md CHANGED
@@ -9,18 +9,18 @@ base_model: Aravindan/gpt2out
9
  datasets:
10
  - generator
11
  model-index:
12
- - name: gpt2coder-instruct
13
  results: []
14
  ---
15
 
16
  <!-- This model card has been generated automatically according to the information the Trainer had access to. You
17
  should probably proofread and complete it, then remove this comment. -->
18
 
19
- # gpt2coder-instruct
20
 
21
  This model is a fine-tuned version of [Aravindan/gpt2out](https://huggingface.co/Aravindan/gpt2out) on the generator dataset.
22
  It achieves the following results on the evaluation set:
23
- - Loss: 2.2879
24
 
25
  ## Model description
26
 
@@ -47,16 +47,14 @@ The following hyperparameters were used during training:
47
  - total_train_batch_size: 80
48
  - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
49
  - lr_scheduler_type: constant
50
- - training_steps: 100
51
  - mixed_precision_training: Native AMP
52
 
53
  ### Training results
54
 
55
  | Training Loss | Epoch | Step | Validation Loss |
56
  |:-------------:|:------:|:----:|:---------------:|
57
- | 2.624 | 0.0147 | 30 | 2.4234 |
58
- | 2.5018 | 0.0294 | 60 | 2.3461 |
59
- | 2.4781 | 0.0441 | 90 | 2.2879 |
60
 
61
 
62
  ### Framework versions
 
9
  datasets:
10
  - generator
11
  model-index:
12
+ - name: output_dir
13
  results: []
14
  ---
15
 
16
  <!-- This model card has been generated automatically according to the information the Trainer had access to. You
17
  should probably proofread and complete it, then remove this comment. -->
18
 
19
+ # output_dir
20
 
21
  This model is a fine-tuned version of [Aravindan/gpt2out](https://huggingface.co/Aravindan/gpt2out) on the generator dataset.
22
  It achieves the following results on the evaluation set:
23
+ - Loss: 2.4231
24
 
25
  ## Model description
26
 
 
47
  - total_train_batch_size: 80
48
  - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
49
  - lr_scheduler_type: constant
50
+ - training_steps: 50
51
  - mixed_precision_training: Native AMP
52
 
53
  ### Training results
54
 
55
  | Training Loss | Epoch | Step | Validation Loss |
56
  |:-------------:|:------:|:----:|:---------------:|
57
+ | 2.6346 | 0.0147 | 30 | 2.4231 |
 
 
58
 
59
 
60
  ### Framework versions
adapter_model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:34ce1a3da80fb0ec2881ee2f099c7cc7de7ce5e529b6eb6d882c412dab2c4232
3
  size 2362376
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:3c9d94f5357649f82fb50b00753f4dea566bf93d56c116eafef6e0d4bd7d819b
3
  size 2362376
runs/Jun13_11-30-16_64efe294a367/events.out.tfevents.1718278359.64efe294a367.34.0 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:30da07901f6b776cbe9693a17874180f77545ed3e8ea3020debb65208a38c02a
3
+ size 6659
training_args.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:4dd68b478a4f85dbd16967cf79e92cf4e7be5afb41510aaffa1744ebf3da8149
3
  size 5368
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:545fdcdfa578c78beef917f63c6d03ca8ab1f74ed19ce0cb794697949e4bb254
3
  size 5368