TSebbag commited on
Commit
27a9d02
·
verified ·
1 Parent(s): 5dadda1

update evaluation

Browse files
Files changed (1) hide show
  1. README.md +15 -0
README.md CHANGED
@@ -15,3 +15,18 @@ tags:
15
 
16
  [AdminBERT-16GB](example) is a French language model adapted on a large corpus of 50 millions French administrative texts. It is a derivative of CamemBERT model, which is based on the RoBERTa architecture. AdminBERT-16GB is trained using the Whole Word Masking (WWM) objective with 30% mask rate for 3 epochs on 24 A100 GPUs. The dataset used for training is [Adminset](https://huggingface.co/datasets/taln-ls2n/Adminset).
17
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
15
 
16
  [AdminBERT-16GB](example) is a French language model adapted on a large corpus of 50 millions French administrative texts. It is a derivative of CamemBERT model, which is based on the RoBERTa architecture. AdminBERT-16GB is trained using the Whole Word Masking (WWM) objective with 30% mask rate for 3 epochs on 24 A100 GPUs. The dataset used for training is [Adminset](https://huggingface.co/datasets/taln-ls2n/Adminset).
17
 
18
+
19
+ ## Evaluation
20
+
21
+ ### Model Performance
22
+
23
+ | Model | P (%) | R (%) | F1 (%) |
24
+ |------------------------|---------|---------|---------|
25
+ | Wikineural-NER FT | 77.49 | 75.40 | 75.70 |
26
+ | NERmemBERT-Large FT | 77.43 | 78.38 | 77.13 |
27
+ | CamemBERT FT | 77.62 | 79.59 | 77.26 |
28
+ | NERmemBERT-Base FT | 77.99 | 79.59 | 78.34 |
29
+ | AdminBERT-NER 4GB | 78.47 | 80.35 | 79.26 |
30
+ | AdminBERT-NER 16GB | 78.79 | 82.07 | 80.11 |
31
+
32
+ To evaluate each model, we performed five runs and averaged the results on the test set of Adminset-NER.