Model Card for Model ID

This model utilizes a Swin Transformer architecture and has undergone supervised fine-tuning on retinal fundus images from the REFUGE challenge dataset. It is specialized in automated analysis of retinal fundus photographs for glaucoma detection. By extracting hierarchical visual features from input fundus images in a cross-scale manner, the model is able to effectively categorize each image as either glaucoma or non-glaucoma. Extensive experiments demonstrate that this model architecture achieves state-of-the-art performance on the REFUGE benchmark for fundus image-based glaucoma classification. To obtain optimal predictions, it is recommended to provide this model with standardized retinal fundus photographs captured using typical fundus imaging protocols.

Model Details

Model Description

  • Developed by: Xu Sun
  • Shared by: Xu Sun
  • Model type: Image classification
  • License: Apache-2.0

Uses

The pretrained model provides glaucoma classification functionality solely based on analysis of retinal fundus images. You may directly utilize the raw model without modification to categorize fundus images as either glaucoma or non-glaucoma. This model is specialized in extracting discriminative features from fundus images to identify glaucoma manifestations. However, to achieve optimal performance, it is highly recommended to fine-tune the model on a representative fundus image dataset prior to deployment in real-world applications.

Bias, Risks, and Limitations

The model is specialized in analyzing retinal fundus images, and is trained exclusively on fundus image datasets to classify images as glaucoma or non-glaucoma. Therefore, to obtain accurate predictions, it is crucial to only input fundus images when using this model. Feeding other types of images may lead to meaningless results. In summary, this model expects fundus images as input for glaucoma classification. For the best performance, please adhere strictly to this input specification.

How to Get Started with the Model

Use the code below to get started with the model.

import cv2
import torch

from transformers import AutoImageProcessor, Swinv2ForImageClassification

image = cv2.imread('./example.jpg')
image = cv2.cvtColor(image, cv2.COLOR_BGR2RGB)

processor = AutoImageProcessor.from_pretrained("pamixsun/swinv2_tiny_for_glaucoma_classification")
model = Swinv2ForImageClassification.from_pretrained("pamixsun/swinv2_tiny_for_glaucoma_classification")

inputs = processor(image, return_tensors="pt")

with torch.no_grad():
    logits = model(**inputs).logits

# model predicts either glaucoma or non-glaucoma.
predicted_label = logits.argmax(-1).item()
print(model.config.id2label[predicted_label])

Citation [optional]

BibTeX:

[More Information Needed]

APA:

[More Information Needed]

Model Card Contact

Downloads last month
91
Safetensors
Model size
27.6M params
Tensor type
F32
Β·
Inference Examples
This model does not have enough activity to be deployed to Inference API (serverless) yet. Increase its social visibility and check back later, or deploy to Inference Endpoints (dedicated) instead.

Spaces using pamixsun/swinv2_tiny_for_glaucoma_classification 5