abspython commited on
Commit
42d7250
·
1 Parent(s): 8c4f363
README.md CHANGED
@@ -2,7 +2,5 @@
2
  language: en
3
  license: other
4
  ---
5
-
6
- # DistilBERT finetuned
7
-
8
- This model is a fine-tune checkpoint of [DistilBERT-base-uncased](https://huggingface.co/distilbert-base-uncased).
 
2
  language: en
3
  license: other
4
  ---
5
+ TDistilBERT finetuned
6
+ This model is a fine-tune checkpoint of DistilBERT-base-uncased[https://huggingface.co/distilbert-base-uncased]
 
 
config.json CHANGED
@@ -1,4 +1,5 @@
1
  {
 
2
  "activation": "gelu",
3
  "architectures": [
4
  "DistilBertForSequenceClassification"
@@ -6,7 +7,7 @@
6
  "attention_dropout": 0.1,
7
  "dim": 768,
8
  "dropout": 0.1,
9
- "finetuning_task": "sst-2",
10
  "hidden_dim": 3072,
11
  "id2label": {
12
  "0": "NEGATIVE",
@@ -23,9 +24,12 @@
23
  "n_layers": 6,
24
  "output_past": true,
25
  "pad_token_id": 0,
 
26
  "qa_dropout": 0.1,
27
  "seq_classif_dropout": 0.2,
28
  "sinusoidal_pos_embds": false,
29
  "tie_weights_": true,
30
- "vocab_size": 30522
 
 
31
  }
 
1
  {
2
+ "_name_or_path": "./",
3
  "activation": "gelu",
4
  "architectures": [
5
  "DistilBertForSequenceClassification"
 
7
  "attention_dropout": 0.1,
8
  "dim": 768,
9
  "dropout": 0.1,
10
+ "finetuning_task": "sst2",
11
  "hidden_dim": 3072,
12
  "id2label": {
13
  "0": "NEGATIVE",
 
24
  "n_layers": 6,
25
  "output_past": true,
26
  "pad_token_id": 0,
27
+ "problem_type": "single_label_classification",
28
  "qa_dropout": 0.1,
29
  "seq_classif_dropout": 0.2,
30
  "sinusoidal_pos_embds": false,
31
  "tie_weights_": true,
32
+ "torch_dtype": "float32",
33
+ "transformers_version": "4.11.0.dev0",
34
+ "vocab_size": 28996
35
  }
convert_flax_to_pytorch.py ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ from transformers import AutoModelForSequenceClassification
2
+ model = AutoModelForSequenceClassification.from_pretrained("./", from_flax=True)
3
+ model.save_pretrained("./")
convert_pytorch_to_flax.py ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ from transformers import FlaxAutoModelForSequenceClassification
2
+ model = FlaxAutoModelForSequenceClassification.from_pretrained("./", from_pt=True)
3
+ model.save_pretrained("./")
convert_pytorch_to_tensorflow.py ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ from transformers import TFAutoModelForSequenceClassification
2
+ model = TFAutoModelForSequenceClassification.from_pretrained("./", from_pt=True)
3
+ model.save_pretrained("./")
rust_model.ot → flax_model.msgpack RENAMED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:9db97da21b97a5e6db1212ce6a810a0c5e22c99daefe3355bae2117f78a0abb9
3
- size 267846324
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:f3be41725be41336d92d410cbeae93b4dc2a34648b232e5b4ecd1526952e6fd3
3
+ size 263135829
pytorch_model.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:60554cbd7781b09d87f1ececbea8c064b94e49a7f03fd88e8775bfe6cc3d9f88
3
- size 267844284
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:dd2a7a30f0822b2d5d6a4b1ead1ceafea2cc15867fc3650b7f5348e64f3818e0
3
+ size 263170801
special_tokens_map.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"unk_token": "[UNK]", "sep_token": "[SEP]", "pad_token": "[PAD]", "cls_token": "[CLS]", "mask_token": "[MASK]"}
tf_model.h5 CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:b44df675bb34ccd8e57c14292c811ac7358b7c8e37c7f212745f640cd6019ac8
3
- size 267949840
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:f4bc3ec52b550e50fd957d5ac563dcbc3bc10a854044b447018711510b20e2ac
3
+ size 263264024
tokenizer.json ADDED
The diff for this file is too large to render. See raw diff
 
tokenizer_config.json CHANGED
@@ -1 +1 @@
1
- {"model_max_length": 512, "do_lower_case": true}
 
1
+ {"do_lower_case": false, "unk_token": "[UNK]", "sep_token": "[SEP]", "pad_token": "[PAD]", "cls_token": "[CLS]", "mask_token": "[MASK]", "tokenize_chinese_chars": true, "strip_accents": null, "model_max_length": 512, "special_tokens_map_file": null, "name_or_path": "/tmp/sst2", "tokenizer_class": "DistilBertTokenizer"}
vocab.txt CHANGED
The diff for this file is too large to render. See raw diff