rwang5688 commited on
Commit
5986030
1 Parent(s): 96da8cc

Training in progress, epoch 1

Browse files
config.json CHANGED
@@ -14,10 +14,12 @@
14
  "n_heads": 12,
15
  "n_layers": 6,
16
  "pad_token_id": 0,
 
17
  "qa_dropout": 0.1,
18
  "seq_classif_dropout": 0.2,
19
  "sinusoidal_pos_embds": false,
20
  "tie_weights_": true,
21
- "transformers_version": "4.21.0",
 
22
  "vocab_size": 30522
23
  }
 
14
  "n_heads": 12,
15
  "n_layers": 6,
16
  "pad_token_id": 0,
17
+ "problem_type": "single_label_classification",
18
  "qa_dropout": 0.1,
19
  "seq_classif_dropout": 0.2,
20
  "sinusoidal_pos_embds": false,
21
  "tie_weights_": true,
22
+ "torch_dtype": "float32",
23
+ "transformers_version": "4.12.0",
24
  "vocab_size": 30522
25
  }
pytorch_model.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:acd6d95a2ddf8fe4baa0816d481a67ae939e87a8697e7e5b28e2ae79b94db70d
3
  size 267861719
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:78bd0b0d9e333358b3d0013986de62ba0b1063e6a4faebda12e947886ad9632a
3
  size 267861719
special_tokens_map.json CHANGED
@@ -1,7 +1 @@
1
- {
2
- "cls_token": "[CLS]",
3
- "mask_token": "[MASK]",
4
- "pad_token": "[PAD]",
5
- "sep_token": "[SEP]",
6
- "unk_token": "[UNK]"
7
- }
 
1
+ {"unk_token": "[UNK]", "sep_token": "[SEP]", "pad_token": "[PAD]", "cls_token": "[CLS]", "mask_token": "[MASK]"}
 
 
 
 
 
 
tokenizer.json CHANGED
The diff for this file is too large to render. See raw diff
 
tokenizer_config.json CHANGED
@@ -1,14 +1 @@
1
- {
2
- "cls_token": "[CLS]",
3
- "do_lower_case": true,
4
- "mask_token": "[MASK]",
5
- "model_max_length": 512,
6
- "name_or_path": "distilbert-base-uncased",
7
- "pad_token": "[PAD]",
8
- "sep_token": "[SEP]",
9
- "special_tokens_map_file": null,
10
- "strip_accents": null,
11
- "tokenize_chinese_chars": true,
12
- "tokenizer_class": "DistilBertTokenizer",
13
- "unk_token": "[UNK]"
14
- }
 
1
+ {"do_lower_case": true, "unk_token": "[UNK]", "sep_token": "[SEP]", "pad_token": "[PAD]", "cls_token": "[CLS]", "mask_token": "[MASK]", "tokenize_chinese_chars": true, "strip_accents": null, "model_max_length": 512, "special_tokens_map_file": null, "name_or_path": "distilbert-base-uncased", "tokenizer_class": "DistilBertTokenizer"}
 
 
 
 
 
 
 
 
 
 
 
 
 
training_args.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:240313fdad20e2e89395e7dae197c0efbacd8700c61dc8ef559a8469a2af2d4f
3
  size 2863
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:252f2e3714f5c40d88062391d2b7ffe5340239325fcf274c7974615d00d98b84
3
  size 2863