Upload 3 files
Browse files- .gitattributes +1 -0
- config.json +1 -0
- model.pt +3 -0
- precomputed_entity_descriptions_emb_wikipedia_6269457-300.np +3 -0
.gitattributes
CHANGED
@@ -33,3 +33,4 @@ saved_model/**/* filter=lfs diff=lfs merge=lfs -text
|
|
33 |
*.zip filter=lfs diff=lfs merge=lfs -text
|
34 |
*.zst filter=lfs diff=lfs merge=lfs -text
|
35 |
*tfevents* filter=lfs diff=lfs merge=lfs -text
|
|
|
|
33 |
*.zip filter=lfs diff=lfs merge=lfs -text
|
34 |
*.zst filter=lfs diff=lfs merge=lfs -text
|
35 |
*tfevents* filter=lfs diff=lfs merge=lfs -text
|
36 |
+
precomputed_entity_descriptions_emb_wikipedia_6269457-300.np filter=lfs diff=lfs merge=lfs -text
|
config.json
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
{"data_dir": "./data", "transformer_name": "roberta-base", "max_seq": 510, "learning_rate": 5e-05, "num_train_epochs": 2, "freeze_all_bert_layers": false, "gradient_accumulation_steps": 1, "per_gpu_batch_size": 12, "freeze_embedding_layers": false, "freeze_layers": [], "n_gpu": 4, "lr_ner_scale": 100, "ner_layer_dropout": 0.1, "ed_layer_dropout": 0.05, "max_candidates": 30, "warmup_steps": 10000, "logging_steps": 500, "save_steps": 500, "detach_ed_layer": true, "only_ner": false, "only_ed": false, "md_layer_dropout": 0.1, "debug": false, "beta_desc": 0.1, "beta_et": 0.0, "beta_ed": 0.0, "sep_token_id": 2, "cls_token_id": 0, "mask_token_id": 50264, "pad_token_id": 1, "vocab_size": 50265, "ner_tag_to_ix": {"O": 0, "B-MENTION": 1, "I-MENTION": 2}}
|
model.pt
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:20d8bd8dde4471a3a941b9796b22f1e77e8cb4bc08c090e4cb5549fe2901049e
|
3 |
+
size 725973509
|
precomputed_entity_descriptions_emb_wikipedia_6269457-300.np
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:b295cbf1a0c2025bbe547c320d1055dff21f8dabf8c3d684b0f16dabbbe6b43c
|
3 |
+
size 3761674200
|