Clinical Text Classification Model

Model Description

Clinical text classification model with fine-tuned contrastive encoder

This model is trained for clinical text classification with the following labels: Absent, Hypothetical, Present

Model Architecture

  • Base model: Clinical ModernBERT with contrastive learning
  • Classification head: 2-layer neural network
  • Dropout rate: 0.1

Usage

from transformers import AutoTokenizer, AutoModelForSequenceClassification
import torch

# Load model and tokenizer
model_name = "nikhil061307/clinical-classifier-finetuned "
tokenizer = AutoTokenizer.from_pretrained(model_name)
model = AutoModelForSequenceClassification.from_pretrained(model_name)

# Example usage
text = "Patient presents with [ENTITY]chest pain[/ENTITY] and shortness of breath."
inputs = tokenizer(text, return_tensors="pt", truncation=True, padding=True, max_length=512)

with torch.no_grad():
    outputs = model(**inputs)
    predictions = torch.nn.functional.softmax(outputs.logits, dim=-1)
    predicted_class = torch.argmax(outputs.logits, dim=-1)

print(f"Predicted class: {predicted_class.item()}")
print(f"Probabilities: {predictions[0].tolist()}")

Label Mapping

{
  "Absent": 0,
  "Hypothetical": 1,
  "Present": 2
}

Training Data

The model was trained on clinical text data with entity mentions marked using [ENTITY] and [/ENTITY] tags.

Performance

Please refer to the training logs and evaluation metrics provided during model development.

Citation

If you use this model, please cite appropriately.

Downloads last month
1
Safetensors
Model size
0.1B params
Tensor type
F32
ยท
Inference Providers NEW
This model isn't deployed by any Inference Provider. ๐Ÿ™‹ Ask for provider support