dreji18 commited on
Commit
067a14e
1 Parent(s): 9b31cb1

updated model

Browse files
Files changed (3) hide show
  1. README.md +43 -0
  2. pytorch_model.bin +1 -1
  3. tokenizer_config.json +1 -1
README.md CHANGED
@@ -1,3 +1,46 @@
1
  ---
2
  license: apache-2.0
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
3
  ---
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
  ---
2
  license: apache-2.0
3
+ language:
4
+ - en
5
+ tags:
6
+ - Token Classification
7
+ co2_eq_emissions: 0.0279399890043426 Kg
8
+ widget:
9
+ - text: "CASE: A 28-year-old previously healthy man presented with a 6-week history of palpitations.
10
+ The symptoms occurred during rest, 2–3 times per week, lasted up to 30 minutes at a time and were associated with dyspnea.
11
+ Except for a grade 2/6 holosystolic tricuspid regurgitation murmur (best heard at the left sternal border with inspiratory accentuation), physical examination yielded unremarkable findings."
12
+ example_title: "example 1"
13
+ - text: "A 63-year-old woman with no known cardiac history presented with a sudden onset of dyspnea requiring intubation and ventilatory support out of hospital.
14
+ She denied preceding symptoms of chest discomfort, palpitations, syncope or infection.
15
+ The patient was afebrile and normotensive, with a sinus tachycardia of 140 beats/min."
16
+ example_title: "example 2"
17
+ - text: "A 48 year-old female presented with vaginal bleeding and abnormal Pap smears.
18
+ Upon diagnosis of invasive non-keratinizing SCC of the cervix, she underwent a radical hysterectomy with salpingo-oophorectomy which demonstrated positive spread to the pelvic lymph nodes and the parametrium.
19
+ Pathological examination revealed that the tumour also extensively involved the lower uterine segment."
20
+ example_title: "example 3"
21
  ---
22
+
23
+ ## About the Model
24
+ An English Named Entity Recognition model, trained on _________ to recognize the bio-medical entities from a given text corpus (case reports etc.). This model was built on top of distilbert-base-uncased
25
+
26
+ - Dataset :
27
+ - Carbon emission : 0.0279399890043426 Kg
28
+ - Training time : 30.16527 minute
29
+ - GPU used : 1 x GeForce RTX 3060 Laptop GPU
30
+
31
+ ## Usage
32
+ The easiest way is to load the inference api from huggingface and second method is through the pipeline object offered by transformers library.
33
+ ```python
34
+ from transformers import pipeline
35
+ from transformers import AutoTokenizer, AutoModelForTokenClassification
36
+
37
+ tokenizer = AutoTokenizer.from_pretrained("d4data/biomedical-all")
38
+ model = AutoModelForTokenClassification.from_pretrained("d4data/biomedical-all")
39
+
40
+ pipe = pipeline("ner", model=model, tokenizer=tokenizer, aggregation_strategy="simple") # pass device=0 if using gpu
41
+ pipe("""The patient reported no recurrence of palpitations at follow-up 6 months after the ablation.""")
42
+ ```
43
+
44
+ ## Author
45
+ This model is part of the Research topic "AI in Biomedical field" conducted by Deepak John Reji, Shaina Raza. If you use this work (code, model or dataset), please star at:
46
+ > repo yet to create, <>
pytorch_model.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:68379ec0b19ee4c0335ffcf30d7e88f30b556c70868801cab4cc1a6d8c73bc2e
3
  size 265814261
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:bd09191cafbd80946b97c301ee5bd56718f94276edff341d15d11d2992f2d699
3
  size 265814261
tokenizer_config.json CHANGED
@@ -1 +1 @@
1
- {"do_lower_case": true, "unk_token": "[UNK]", "sep_token": "[SEP]", "pad_token": "[PAD]", "cls_token": "[CLS]", "mask_token": "[MASK]", "tokenize_chinese_chars": true, "strip_accents": null, "model_max_length": 512, "special_tokens_map_file": null, "name_or_path": "distilbert-base-uncased", "tokenizer_class": "DistilBertTokenizer"}
 
1
+ {"do_lower_case": true, "unk_token": "[UNK]", "sep_token": "[SEP]", "pad_token": "[PAD]", "cls_token": "[CLS]", "mask_token": "[MASK]", "tokenize_chinese_chars": true, "strip_accents": null, "cuda": 0, "model_max_length": 512, "special_tokens_map_file": null, "name_or_path": "distilbert-base-uncased", "tokenizer_class": "DistilBertTokenizer"}