Instructions to use SU-FMI-AI/multiclinner_enigma_cz_procedure_robeczech-os1 with libraries, inference providers, notebooks, and local apps. Follow these links to get started.
- Libraries
- Transformers
How to use SU-FMI-AI/multiclinner_enigma_cz_procedure_robeczech-os1 with Transformers:
# Use a pipeline as a high-level helper from transformers import pipeline pipe = pipeline("token-classification", model="SU-FMI-AI/multiclinner_enigma_cz_procedure_robeczech-os1")# Load model directly from transformers import AutoTokenizer, AutoModelForTokenClassification tokenizer = AutoTokenizer.from_pretrained("SU-FMI-AI/multiclinner_enigma_cz_procedure_robeczech-os1") model = AutoModelForTokenClassification.from_pretrained("SU-FMI-AI/multiclinner_enigma_cz_procedure_robeczech-os1") - Notebooks
- Google Colab
- Kaggle
multiclinner_enigma_cz_procedure_robeczech-os1
Czech clinical Named Entity Recognition model for the PROCEDURE entity
type, fine-tuned from ufal/robeczech-base on the
Czech portion of the MultiClinAI 2026 IberLEF shared task.
Developed by Team Enigma at the Faculty of Mathematics and Informatics,
Sofia University.
Summary
| Task | Token classification (BIO), single entity type |
| Entity type | PROCEDURE (medical procedures and interventions) |
| Language | Czech (cs) |
| Base model | ufal/robeczech-base |
| Architecture | Transformer (softmax) |
| Augmentation | Morphological synonyms + 1x oversample of entity-bearing docs |
| Training data | MultiClinAI Czech train + dev combined (1,258 documents) plus augmentation |
| Test F1 (strict) | 0.6620 (char F1 0.7972) |
Quick start
from transformers import AutoTokenizer, AutoModelForTokenClassification, pipeline
repo = "SU-FMI-AI/multiclinner_enigma_cz_procedure_robeczech-os1"
tokenizer = AutoTokenizer.from_pretrained(repo)
model = AutoModelForTokenClassification.from_pretrained(repo)
ner = pipeline(
"token-classification",
model=model,
tokenizer=tokenizer,
aggregation_strategy="simple",
)
text = "Pacient byl přijat s hypertenzà a podstoupil koronarografii."
for span in ner(text):
print(span)
Each predicted span is a dictionary with keys entity_group,
score, word, start, end.
Label set
The model predicts a single entity type using the BIO tagging scheme:
| ID | Label | Meaning |
|---|---|---|
| 0 | O |
outside any entity |
| 1 | B-PROCEDURE |
beginning of a PROCEDURE mention |
| 2 | I-PROCEDURE |
inside a PROCEDURE mention |
Intended use
- Extracting
PROCEDUREmentions from Czech clinical text (discharge summaries, case reports, medical records). - Building block for ensembles (this model was deployed as part of an ensemble in the original submission).
- A starting point for further fine-tuning on related Czech biomedical corpora.
Out-of-scope use
- Other languages. Although the XLM-RoBERTa based variants share a multilingual encoder, the classification head was only trained on Czech.
- Other domains. The model was not exposed to non-clinical text, social media or layperson descriptions.
- Clinical decision making. This is a research artifact. Do not use it as the sole input to any clinical decision.
Training data
- Source. MultiClinAI Czech NER: 1,006 train documents
and 252 dev documents per entity type in BRAT standoff format, derived
from the DisTEMIST, SympTEMIST and MedProcNER corpora translated and
annotation-projected to Czech. For the final submission models the gold
train + devsets are merged into a single training partition (no held-out validation). - Augmentation. Morphological synonym replacement + oversampling. Morphological augmentation as described above, with one additional copy of every training document that contains at least one entity annotation, added before the augmented documents are appended. This increases the relative weight of gold supervision over the synthetic data.
- Tokenisation. SentencePiece tokenizer inherited from the base model.
Training procedure
| Hyperparameter | Value |
|---|---|
| Base model | ufal/robeczech-base |
| Head | SOFTMAX |
| Optimiser | AdamW |
| Learning rate | 2e-5 |
| Batch size | 64 |
| Epochs | 10 |
| Max sequence length | 512 |
| Input granularity | Sentence-level |
| Warmup ratio | 0.1 |
| Weight decay | 0.01 |
| Seed | 42 |
| Mixed precision | fp16 (CUDA) |
Token classification head: backbone hidden states are projected through a linear layer to BIO logits and decoded greedily (argmax per token).
Evaluation
Held-out development set
Best dev-set entity-level F1 observed during development: 0.735.
MultiClinAI Czech, official blind test set
Run name in the official MultiClinAI ranking: robeczech-os1_cz_procedure.
| Metric | Strict | Character-level |
|---|---|---|
| Precision | 0.6644 | 0.8021 |
| Recall | 0.6596 | 0.7922 |
| F1 | 0.6620 | 0.7972 |
Strict matching requires the predicted span to exactly match a gold span (same start, end, and type). Character-level matching gives partial credit for overlapping spans.
Related models
Other models for the same entity type:
SU-FMI-AI/multiclinner_enigma_cz_procedure_xlmr-crf: xlm-roberta-base, Morphological synonyms (curated + Wikidata), CRF head, test F1 = 0.6552.
License
Released under the apache-2.0 license. Base-model and dataset licenses apply to their respective artifacts.
Code and resources
Training code, augmentation pipeline, ablation log and evaluation scripts are available in the project's GitHub repository: https://github.com/TeogopK/MultiClinAI-Czech.
- Downloads last month
- 18
Model tree for SU-FMI-AI/multiclinner_enigma_cz_procedure_robeczech-os1
Base model
ufal/robeczech-baseEvaluation results
- Strict Precision on MultiClinAI Czech (PROCEDURE)self-reported0.664
- Strict Recall on MultiClinAI Czech (PROCEDURE)self-reported0.660
- Strict F1 on MultiClinAI Czech (PROCEDURE)self-reported0.662
- Char-level Precision on MultiClinAI Czech (PROCEDURE)self-reported0.802
- Char-level Recall on MultiClinAI Czech (PROCEDURE)self-reported0.792
- Char-level F1 on MultiClinAI Czech (PROCEDURE)self-reported0.797