ndiy commited on
Commit
206751d
·
verified ·
1 Parent(s): 95c1496

End of training

Browse files
README.md CHANGED
@@ -1,46 +1,68 @@
1
- ---
2
- tags:
3
- - generated_from_trainer
4
- model-index:
5
- - name: ASAP_SENT
6
- results: []
7
- ---
8
-
9
- <!-- This model card has been generated automatically according to the information the Trainer had access to. You
10
- should probably proofread and complete it, then remove this comment. -->
11
-
12
- # ASAP_SENT
13
-
14
- This model was trained from scratch on the None dataset.
15
-
16
- ## Model description
17
-
18
- More information needed
19
-
20
- ## Intended uses & limitations
21
-
22
- More information needed
23
-
24
- ## Training and evaluation data
25
-
26
- More information needed
27
-
28
- ## Training procedure
29
-
30
- ### Training hyperparameters
31
-
32
- The following hyperparameters were used during training:
33
- - learning_rate: 2e-05
34
- - train_batch_size: 64
35
- - eval_batch_size: 16
36
- - seed: 42
37
- - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
38
- - lr_scheduler_type: linear
39
- - num_epochs: 3
40
-
41
- ### Framework versions
42
-
43
- - Transformers 4.40.2
44
- - Pytorch 2.3.0+cu121
45
- - Datasets 2.19.1
46
- - Tokenizers 0.19.1
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ base_model: techthiyanes/chinese_sentiment
3
+ tags:
4
+ - generated_from_trainer
5
+ metrics:
6
+ - accuracy
7
+ model-index:
8
+ - name: ASAP_SENT
9
+ results: []
10
+ ---
11
+
12
+ <!-- This model card has been generated automatically according to the information the Trainer had access to. You
13
+ should probably proofread and complete it, then remove this comment. -->
14
+
15
+ # ASAP_SENT
16
+
17
+ This model is a fine-tuned version of [techthiyanes/chinese_sentiment](https://huggingface.co/techthiyanes/chinese_sentiment) on an unknown dataset.
18
+ It achieves the following results on the evaluation set:
19
+ - Loss: 1.8509
20
+ - Accuracy: 0.5615
21
+
22
+ ## Model description
23
+
24
+ More information needed
25
+
26
+ ## Intended uses & limitations
27
+
28
+ More information needed
29
+
30
+ ## Training and evaluation data
31
+
32
+ More information needed
33
+
34
+ ## Training procedure
35
+
36
+ ### Training hyperparameters
37
+
38
+ The following hyperparameters were used during training:
39
+ - learning_rate: 3e-05
40
+ - train_batch_size: 32
41
+ - eval_batch_size: 8
42
+ - seed: 42
43
+ - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
44
+ - lr_scheduler_type: linear
45
+ - num_epochs: 10
46
+
47
+ ### Training results
48
+
49
+ | Training Loss | Epoch | Step | Validation Loss | Accuracy |
50
+ |:-------------:|:-----:|:-----:|:---------------:|:--------:|
51
+ | 0.9289 | 1.0 | 1152 | 0.8937 | 0.5879 |
52
+ | 0.8395 | 2.0 | 2304 | 0.9166 | 0.5913 |
53
+ | 0.7698 | 3.0 | 3456 | 0.9561 | 0.5879 |
54
+ | 0.6523 | 4.0 | 4608 | 1.0678 | 0.5787 |
55
+ | 0.4581 | 5.0 | 5760 | 1.2595 | 0.5648 |
56
+ | 0.3461 | 6.0 | 6912 | 1.4034 | 0.5595 |
57
+ | 0.2374 | 7.0 | 8064 | 1.5402 | 0.5630 |
58
+ | 0.2438 | 8.0 | 9216 | 1.7174 | 0.5607 |
59
+ | 0.2324 | 9.0 | 10368 | 1.7918 | 0.5640 |
60
+ | 0.1555 | 10.0 | 11520 | 1.8509 | 0.5615 |
61
+
62
+
63
+ ### Framework versions
64
+
65
+ - Transformers 4.39.3
66
+ - Pytorch 2.1.2
67
+ - Datasets 2.18.0
68
+ - Tokenizers 0.15.2
config.json CHANGED
@@ -1,39 +1,41 @@
1
- {
2
- "_name_or_path": "techthiyanes/chinese_sentiment",
3
- "architectures": [
4
- "BertForSequenceClassification"
5
- ],
6
- "attention_probs_dropout_prob": 0.1,
7
- "classifier_dropout": null,
8
- "gradient_checkpointing": false,
9
- "hidden_act": "gelu",
10
- "hidden_dropout_prob": 0.1,
11
- "hidden_size": 768,
12
- "id2label": {
13
- "0": "1",
14
- "1": "2",
15
- "2": "3",
16
- "3": "4",
17
- "4": "5"
18
- },
19
- "initializer_range": 0.02,
20
- "intermediate_size": 3072,
21
- "label2id": {
22
- "1": 0,
23
- "2": 1,
24
- "3": 2,
25
- "4": 3,
26
- "5": 4
27
- },
28
- "layer_norm_eps": 1e-12,
29
- "max_position_embeddings": 512,
30
- "model_type": "bert",
31
- "num_attention_heads": 12,
32
- "num_hidden_layers": 12,
33
- "pad_token_id": 0,
34
- "position_embedding_type": "absolute",
35
- "transformers_version": "4.40.2",
36
- "type_vocab_size": 2,
37
- "use_cache": true,
38
- "vocab_size": 21128
39
- }
 
 
 
1
+ {
2
+ "_name_or_path": "techthiyanes/chinese_sentiment",
3
+ "architectures": [
4
+ "BertForSequenceClassification"
5
+ ],
6
+ "attention_probs_dropout_prob": 0.1,
7
+ "classifier_dropout": null,
8
+ "gradient_checkpointing": false,
9
+ "hidden_act": "gelu",
10
+ "hidden_dropout_prob": 0.1,
11
+ "hidden_size": 768,
12
+ "id2label": {
13
+ "0": "star 1",
14
+ "1": "star 2",
15
+ "2": "star 3",
16
+ "3": "star 4",
17
+ "4": "star 5"
18
+ },
19
+ "initializer_range": 0.02,
20
+ "intermediate_size": 3072,
21
+ "label2id": {
22
+ "star 1": 0,
23
+ "star 2": 1,
24
+ "star 3": 2,
25
+ "star 4": 3,
26
+ "star 5": 4
27
+ },
28
+ "layer_norm_eps": 1e-12,
29
+ "max_position_embeddings": 512,
30
+ "model_type": "bert",
31
+ "num_attention_heads": 12,
32
+ "num_hidden_layers": 12,
33
+ "pad_token_id": 0,
34
+ "position_embedding_type": "absolute",
35
+ "problem_type": "single_label_classification",
36
+ "torch_dtype": "float32",
37
+ "transformers_version": "4.39.3",
38
+ "type_vocab_size": 2,
39
+ "use_cache": true,
40
+ "vocab_size": 21128
41
+ }
model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:45dfc32ef0edd8709ba395d8559ed5481727ecf319d0910e3292ffe24c72acac
3
- size 409109652
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:db77d7a2393f755120c6eb0c4c76d883a36cf2c8c6a8307c2770a828eebc83f2
3
+ size 409109468
runs/May09_14-46-59_d8617a69d045/events.out.tfevents.1715266038.d8617a69d045.25.0 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:fbcef00c4f526364053eb0e1ad223316c9d2ce8761f91af12cc74a545ce8ac6e
3
+ size 251440
special_tokens_map.json CHANGED
@@ -1,7 +1,7 @@
1
- {
2
- "cls_token": "[CLS]",
3
- "mask_token": "[MASK]",
4
- "pad_token": "[PAD]",
5
- "sep_token": "[SEP]",
6
- "unk_token": "[UNK]"
7
- }
 
1
+ {
2
+ "cls_token": "[CLS]",
3
+ "mask_token": "[MASK]",
4
+ "pad_token": "[PAD]",
5
+ "sep_token": "[SEP]",
6
+ "unk_token": "[UNK]"
7
+ }
tokenizer.json CHANGED
@@ -1,7 +1,21 @@
1
  {
2
  "version": "1.0",
3
- "truncation": null,
4
- "padding": null,
 
 
 
 
 
 
 
 
 
 
 
 
 
 
5
  "added_tokens": [
6
  {
7
  "id": 0,
 
1
  {
2
  "version": "1.0",
3
+ "truncation": {
4
+ "direction": "Right",
5
+ "max_length": 200,
6
+ "strategy": "LongestFirst",
7
+ "stride": 0
8
+ },
9
+ "padding": {
10
+ "strategy": {
11
+ "Fixed": 200
12
+ },
13
+ "direction": "Right",
14
+ "pad_to_multiple_of": null,
15
+ "pad_id": 0,
16
+ "pad_type_id": 0,
17
+ "pad_token": "[PAD]"
18
+ },
19
  "added_tokens": [
20
  {
21
  "id": 0,
tokenizer_config.json CHANGED
@@ -1,57 +1,57 @@
1
- {
2
- "added_tokens_decoder": {
3
- "0": {
4
- "content": "[PAD]",
5
- "lstrip": false,
6
- "normalized": false,
7
- "rstrip": false,
8
- "single_word": false,
9
- "special": true
10
- },
11
- "100": {
12
- "content": "[UNK]",
13
- "lstrip": false,
14
- "normalized": false,
15
- "rstrip": false,
16
- "single_word": false,
17
- "special": true
18
- },
19
- "101": {
20
- "content": "[CLS]",
21
- "lstrip": false,
22
- "normalized": false,
23
- "rstrip": false,
24
- "single_word": false,
25
- "special": true
26
- },
27
- "102": {
28
- "content": "[SEP]",
29
- "lstrip": false,
30
- "normalized": false,
31
- "rstrip": false,
32
- "single_word": false,
33
- "special": true
34
- },
35
- "103": {
36
- "content": "[MASK]",
37
- "lstrip": false,
38
- "normalized": false,
39
- "rstrip": false,
40
- "single_word": false,
41
- "special": true
42
- }
43
- },
44
- "clean_up_tokenization_spaces": true,
45
- "cls_token": "[CLS]",
46
- "do_basic_tokenize": true,
47
- "do_lower_case": true,
48
- "mask_token": "[MASK]",
49
- "model_max_length": 1000000000000000019884624838656,
50
- "never_split": null,
51
- "pad_token": "[PAD]",
52
- "sep_token": "[SEP]",
53
- "strip_accents": null,
54
- "tokenize_chinese_chars": true,
55
- "tokenizer_class": "BertTokenizer",
56
- "unk_token": "[UNK]"
57
- }
 
1
+ {
2
+ "added_tokens_decoder": {
3
+ "0": {
4
+ "content": "[PAD]",
5
+ "lstrip": false,
6
+ "normalized": false,
7
+ "rstrip": false,
8
+ "single_word": false,
9
+ "special": true
10
+ },
11
+ "100": {
12
+ "content": "[UNK]",
13
+ "lstrip": false,
14
+ "normalized": false,
15
+ "rstrip": false,
16
+ "single_word": false,
17
+ "special": true
18
+ },
19
+ "101": {
20
+ "content": "[CLS]",
21
+ "lstrip": false,
22
+ "normalized": false,
23
+ "rstrip": false,
24
+ "single_word": false,
25
+ "special": true
26
+ },
27
+ "102": {
28
+ "content": "[SEP]",
29
+ "lstrip": false,
30
+ "normalized": false,
31
+ "rstrip": false,
32
+ "single_word": false,
33
+ "special": true
34
+ },
35
+ "103": {
36
+ "content": "[MASK]",
37
+ "lstrip": false,
38
+ "normalized": false,
39
+ "rstrip": false,
40
+ "single_word": false,
41
+ "special": true
42
+ }
43
+ },
44
+ "clean_up_tokenization_spaces": true,
45
+ "cls_token": "[CLS]",
46
+ "do_basic_tokenize": true,
47
+ "do_lower_case": true,
48
+ "mask_token": "[MASK]",
49
+ "model_max_length": 1000000000000000019884624838656,
50
+ "never_split": null,
51
+ "pad_token": "[PAD]",
52
+ "sep_token": "[SEP]",
53
+ "strip_accents": null,
54
+ "tokenize_chinese_chars": true,
55
+ "tokenizer_class": "BertTokenizer",
56
+ "unk_token": "[UNK]"
57
+ }
training_args.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:28d38072351e4f2a43faed6c11c14b4738328a57def5faa346cd1b01a752fdb4
3
- size 4984
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:cfcfc1556d20ea68c44d324733f77997a69529c4097a8b388d7408d010a21399
3
+ size 4920