palsp commited on
Commit
471ede1
·
verified ·
1 Parent(s): e9a1704

End of training

Browse files
README.md ADDED
@@ -0,0 +1,71 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ base_model: gpt2
3
+ library_name: peft
4
+ license: mit
5
+ tags:
6
+ - generated_from_trainer
7
+ model-index:
8
+ - name: gpt2-lora
9
+ results: []
10
+ ---
11
+
12
+ <!-- This model card has been generated automatically according to the information the Trainer had access to. You
13
+ should probably proofread and complete it, then remove this comment. -->
14
+
15
+ # gpt2-lora
16
+
17
+ This model is a fine-tuned version of [gpt2](https://huggingface.co/gpt2) on an unknown dataset.
18
+ It achieves the following results on the evaluation set:
19
+ - Loss: 2.2244
20
+
21
+ ## Model description
22
+
23
+ More information needed
24
+
25
+ ## Intended uses & limitations
26
+
27
+ More information needed
28
+
29
+ ## Training and evaluation data
30
+
31
+ More information needed
32
+
33
+ ## Training procedure
34
+
35
+ ### Training hyperparameters
36
+
37
+ The following hyperparameters were used during training:
38
+ - learning_rate: 0.0005
39
+ - train_batch_size: 4
40
+ - eval_batch_size: 4
41
+ - seed: 42
42
+ - gradient_accumulation_steps: 8
43
+ - total_train_batch_size: 32
44
+ - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
45
+ - lr_scheduler_type: cosine
46
+ - num_epochs: 5
47
+ - mixed_precision_training: Native AMP
48
+
49
+ ### Training results
50
+
51
+ | Training Loss | Epoch | Step | Validation Loss |
52
+ |:-------------:|:------:|:----:|:---------------:|
53
+ | 2.352 | 0.4737 | 200 | 2.2613 |
54
+ | 2.363 | 0.9473 | 400 | 2.2484 |
55
+ | 2.3263 | 1.4210 | 600 | 2.2433 |
56
+ | 2.3454 | 1.8946 | 800 | 2.2371 |
57
+ | 2.3287 | 2.3683 | 1000 | 2.2321 |
58
+ | 2.3003 | 2.8419 | 1200 | 2.2296 |
59
+ | 2.2973 | 3.3156 | 1400 | 2.2266 |
60
+ | 2.296 | 3.7892 | 1600 | 2.2257 |
61
+ | 2.2924 | 4.2629 | 1800 | 2.2253 |
62
+ | 2.2923 | 4.7365 | 2000 | 2.2244 |
63
+
64
+
65
+ ### Framework versions
66
+
67
+ - PEFT 0.12.0
68
+ - Transformers 4.42.4
69
+ - Pytorch 2.4.0+cu121
70
+ - Datasets 2.21.0
71
+ - Tokenizers 0.19.1
adapter_config.json CHANGED
@@ -3,7 +3,7 @@
3
  "auto_mapping": null,
4
  "base_model_name_or_path": "gpt2",
5
  "bias": "none",
6
- "fan_in_fan_out": true,
7
  "inference_mode": true,
8
  "init_lora_weights": true,
9
  "layer_replication": null,
 
3
  "auto_mapping": null,
4
  "base_model_name_or_path": "gpt2",
5
  "bias": "none",
6
+ "fan_in_fan_out": false,
7
  "inference_mode": true,
8
  "init_lora_weights": true,
9
  "layer_replication": null,
runs/Aug28_14-05-32_86d6d300df19/events.out.tfevents.1724853945.86d6d300df19.215.1 CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:d7a9ef01f8e0fb7c8bcc1cc84d391424ad7bf57b82cab4a9f8bf9b9cf645ede8
3
- size 12197
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:6fa42912ed4957c68a35aa22fb005c26d225dd54efb19d8ffe176490864e1fae
3
+ size 12551
runs/Aug28_14-47-01_86d6d300df19/events.out.tfevents.1724856426.86d6d300df19.215.2 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:096980cc5bbb1603d08160e992b72ad01a87d088f60795877a6e57588b65d2c5
3
+ size 5414
special_tokens_map.json CHANGED
@@ -1,6 +1,30 @@
1
  {
2
- "bos_token": "<|endoftext|>",
3
- "eos_token": "<|endoftext|>",
4
- "pad_token": "<|endoftext|>",
5
- "unk_token": "<|endoftext|>"
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
6
  }
 
1
  {
2
+ "bos_token": {
3
+ "content": "<|endoftext|>",
4
+ "lstrip": false,
5
+ "normalized": true,
6
+ "rstrip": false,
7
+ "single_word": false
8
+ },
9
+ "eos_token": {
10
+ "content": "<|endoftext|>",
11
+ "lstrip": false,
12
+ "normalized": true,
13
+ "rstrip": false,
14
+ "single_word": false
15
+ },
16
+ "pad_token": {
17
+ "content": "<|endoftext|>",
18
+ "lstrip": false,
19
+ "normalized": true,
20
+ "rstrip": false,
21
+ "single_word": false
22
+ },
23
+ "unk_token": {
24
+ "content": "<|endoftext|>",
25
+ "lstrip": false,
26
+ "normalized": true,
27
+ "rstrip": false,
28
+ "single_word": false
29
+ }
30
  }
tokenizer_config.json CHANGED
@@ -13,8 +13,12 @@
13
  "bos_token": "<|endoftext|>",
14
  "clean_up_tokenization_spaces": true,
15
  "eos_token": "<|endoftext|>",
 
16
  "model_max_length": 1024,
17
  "pad_token": "<|endoftext|>",
 
18
  "tokenizer_class": "GPT2Tokenizer",
 
 
19
  "unk_token": "<|endoftext|>"
20
  }
 
13
  "bos_token": "<|endoftext|>",
14
  "clean_up_tokenization_spaces": true,
15
  "eos_token": "<|endoftext|>",
16
+ "max_length": 1024,
17
  "model_max_length": 1024,
18
  "pad_token": "<|endoftext|>",
19
+ "stride": 0,
20
  "tokenizer_class": "GPT2Tokenizer",
21
+ "truncation_side": "right",
22
+ "truncation_strategy": "longest_first",
23
  "unk_token": "<|endoftext|>"
24
  }
training_args.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:04b0c39548a29e17c455cbaafbcca092bc4a1f63a91871685da502262e44ee7b
3
  size 5112
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:ac808cdaefd151e141733b3dceead8564a92a5eebeaabd181fc0e6eec1269385
3
  size 5112