Update README.md
Browse files
README.md
CHANGED
@@ -7,7 +7,18 @@ language:
|
|
7 |
license: mit
|
8 |
model-index:
|
9 |
- name: en_legal_ner_trf
|
10 |
-
results:
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
11 |
---
|
12 |
Indian Legal Named Entity Recognition: Identifying relevant entities in an Indian legal document
|
13 |
|
@@ -39,6 +50,6 @@ Indian Legal Named Entity Recognition: Identifying relevant entities in an India
|
|
39 |
|
40 |
| Type | Score |
|
41 |
| --- | --- |
|
|
|
42 |
| `Precision` | 91.474 |
|
43 |
| `Recall` | 89.956 |
|
44 |
-
| **`F1-Score`** | **90.709** |
|
|
|
7 |
license: mit
|
8 |
model-index:
|
9 |
- name: en_legal_ner_trf
|
10 |
+
results:
|
11 |
+
- task:
|
12 |
+
type: token-classification # Required. Example: automatic-speech-recognition
|
13 |
+
name: Named Entity Recognition # Optional. Example: Speech Recognition
|
14 |
+
dataset:
|
15 |
+
type: Named Entity Recognition
|
16 |
+
name: Indian Legal NER
|
17 |
+
split: Test # Optional. Example: test
|
18 |
+
metrics:
|
19 |
+
- type: F1-Score # Required. Example: wer. Use metric id from https://hf.co/metrics
|
20 |
+
value: 90.709 # Required. Example: 20.90
|
21 |
+
name: Test F1-Score # Optional. Example: Test WER
|
22 |
---
|
23 |
Indian Legal Named Entity Recognition: Identifying relevant entities in an Indian legal document
|
24 |
|
|
|
50 |
|
51 |
| Type | Score |
|
52 |
| --- | --- |
|
53 |
+
| **F1-Score** | **90.709** |
|
54 |
| `Precision` | 91.474 |
|
55 |
| `Recall` | 89.956 |
|
|