truongnguyenxuanvinh commited on
Commit
5ce9122
·
verified ·
1 Parent(s): 433e822

Upload RobertaForSequenceClassification

Browse files
Files changed (2) hide show
  1. config.json +25 -68
  2. model.safetensors +2 -2
config.json CHANGED
@@ -1,87 +1,44 @@
1
  {
2
- "_name_or_path": "bergum/xtremedistil-l6-h384-go-emotion",
3
  "architectures": [
4
- "BertForSequenceClassification"
5
  ],
6
  "attention_probs_dropout_prob": 0.1,
 
7
  "classifier_dropout": null,
8
- "gradient_checkpointing": false,
9
  "hidden_act": "gelu",
10
  "hidden_dropout_prob": 0.1,
11
- "hidden_size": 384,
12
  "id2label": {
13
- "0": "admiration \ud83d\udc4f",
14
- "1": "amusement \ud83d\ude02",
15
- "2": "anger \ud83d\ude21",
16
- "3": "annoyance \ud83d\ude12",
17
- "4": "approval \ud83d\udc4d",
18
- "5": "caring \ud83e\udd17",
19
- "6": "confusion \ud83d\ude15",
20
- "7": "curiosity \ud83e\udd14",
21
- "8": "desire \ud83d\ude0d",
22
- "9": "disappointment \ud83d\ude1e",
23
- "10": "disapproval \ud83d\udc4e",
24
- "11": "disgust \ud83e\udd2e",
25
- "12": "embarrassment \ud83d\ude33",
26
- "13": "excitement \ud83e\udd29",
27
- "14": "fear \ud83d\ude28",
28
- "15": "gratitude \ud83d\ude4f",
29
- "16": "grief \ud83d\ude22",
30
- "17": "joy \ud83d\ude03",
31
- "18": "love \u2764\ufe0f",
32
- "19": "nervousness \ud83d\ude2c",
33
- "20": "optimism \ud83e\udd1e",
34
- "21": "pride \ud83d\ude0c",
35
- "22": "realization \ud83d\udca1",
36
- "23": "relief \ud83d\ude05",
37
- "24": "remorse \ud83d\ude1e",
38
- "25": "sadness \ud83d\ude1e",
39
- "26": "surprise \ud83d\ude32",
40
- "27": "neutral \ud83d\ude10"
41
  },
42
  "initializer_range": 0.02,
43
- "intermediate_size": 1536,
44
  "label2id": {
45
- "admiration \ud83d\udc4f": 0,
46
- "amusement \ud83d\ude02": 1,
47
- "anger \ud83d\ude21": 2,
48
- "annoyance \ud83d\ude12": 3,
49
- "approval \ud83d\udc4d": 4,
50
- "caring \ud83e\udd17": 5,
51
- "confusion \ud83d\ude15": 6,
52
- "curiosity \ud83e\udd14": 7,
53
- "desire \ud83d\ude0d": 8,
54
- "disappointment \ud83d\ude1e": 9,
55
- "disapproval \ud83d\udc4e": 10,
56
- "disgust \ud83e\udd2e": 11,
57
- "embarrassment \ud83d\ude33": 12,
58
- "excitement \ud83e\udd29": 13,
59
- "fear \ud83d\ude28": 14,
60
- "gratitude \ud83d\ude4f": 15,
61
- "grief \ud83d\ude22": 16,
62
- "joy \ud83d\ude03": 17,
63
- "love \u2764\ufe0f": 18,
64
- "nervousness \ud83d\ude2c": 19,
65
- "neutral \ud83d\ude10": 27,
66
- "optimism \ud83e\udd1e": 20,
67
- "pride \ud83d\ude0c": 21,
68
- "realization \ud83d\udca1": 22,
69
- "relief \ud83d\ude05": 23,
70
- "remorse \ud83d\ude1e": 24,
71
- "sadness \ud83d\ude1e": 25,
72
- "surprise \ud83d\ude32": 26
73
  },
74
- "layer_norm_eps": 1e-12,
75
- "max_position_embeddings": 512,
76
- "model_type": "bert",
77
  "num_attention_heads": 12,
78
- "num_hidden_layers": 6,
79
- "pad_token_id": 0,
80
  "position_embedding_type": "absolute",
81
  "problem_type": "multi_label_classification",
82
  "torch_dtype": "float32",
83
  "transformers_version": "4.44.2",
84
- "type_vocab_size": 2,
85
  "use_cache": true,
86
- "vocab_size": 30522
87
  }
 
1
  {
2
+ "_name_or_path": "FacebookAI/roberta-base",
3
  "architectures": [
4
+ "RobertaForSequenceClassification"
5
  ],
6
  "attention_probs_dropout_prob": 0.1,
7
+ "bos_token_id": 0,
8
  "classifier_dropout": null,
9
+ "eos_token_id": 2,
10
  "hidden_act": "gelu",
11
  "hidden_dropout_prob": 0.1,
12
+ "hidden_size": 768,
13
  "id2label": {
14
+ "0": "LABEL_0",
15
+ "1": "LABEL_1",
16
+ "2": "LABEL_2",
17
+ "3": "LABEL_3",
18
+ "4": "LABEL_4",
19
+ "5": "LABEL_5"
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
20
  },
21
  "initializer_range": 0.02,
22
+ "intermediate_size": 3072,
23
  "label2id": {
24
+ "LABEL_0": 0,
25
+ "LABEL_1": 1,
26
+ "LABEL_2": 2,
27
+ "LABEL_3": 3,
28
+ "LABEL_4": 4,
29
+ "LABEL_5": 5
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
30
  },
31
+ "layer_norm_eps": 1e-05,
32
+ "max_position_embeddings": 514,
33
+ "model_type": "roberta",
34
  "num_attention_heads": 12,
35
+ "num_hidden_layers": 12,
36
+ "pad_token_id": 1,
37
  "position_embedding_type": "absolute",
38
  "problem_type": "multi_label_classification",
39
  "torch_dtype": "float32",
40
  "transformers_version": "4.44.2",
41
+ "type_vocab_size": 1,
42
  "use_cache": true,
43
+ "vocab_size": 50265
44
  }
model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:4d441dd4a975babd1edd3c2b888b4263025cac424a90f65220257eda99de9ed3
3
- size 90908000
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:fe74a4e75f3e97ff09c97527f88076a03b265bd1c4aac1b3e5b1307a86370b5b
3
+ size 498625128