Edit model card

Model Card for CLIP-RoBERTa Fine-Tuned Model

This model is a fine-tuned version of the CLIP model combining a vision model (openai/clip-vit-base-patch32) with a text model (roberta-base). The model is fine-tuned to better handle image-text matching tasks, utilizing a variety of training strategies for enhanced performance.

Model fine tuned for Project: DermAi-Viz

Model Details

Model Description

This model card describes a Vision-Text Dual Encoder model fine-tuned from the original openai/clip-vit-base-patch32 and roberta-base models. The model is specifically adapted for tasks that involve joint processing of images and textual descriptions, leveraging both the image encoding capabilities of the CLIP model and the language understanding of RoBERTa.

Model Sources

  • Repository: GitHub
  • Demo: [Optional, link to a demo if applicable]

Uses

Direct Use

This model can be directly used for image-text matching tasks, such as searching for relevant images based on a textual query or generating captions for images.

Downstream Use

The model can be further fine-tuned for specific tasks such as image classification with text guidance, visual question answering, or any other task that benefits from multi-modal inputs.

Out-of-Scope Use

This model is not suitable for tasks unrelated to image and text processing. It may also not perform well on non-English texts or images outside the scope of the fine-tuning data.

Bias, Risks, and Limitations

As with most large models, this model may inherit biases present in the pretraining datasets. Users should be cautious when deploying this model in sensitive applications, particularly where fairness and bias are of concern. Additionally, the model's performance might degrade on data that is significantly different from the training data, especially in terms of cultural, linguistic, or visual context.

Recommendations

Users should validate the model's performance on their specific use case, particularly looking for biases in the model's predictions. Additional fine-tuning with a carefully curated dataset may be necessary to mitigate biases. It's also recommended to monitor the model's outputs regularly to ensure its predictions remain relevant and unbiased.

How to Get Started with the Model

from transformers import VisionTextDualEncoderModel, VisionTextDualEncoderProcessor

# Load the model and processor
model = VisionTextDualEncoderModel.from_pretrained("your-username/clip-roberta-finetuned")
processor = VisionTextDualEncoderProcessor.from_pretrained("your-username/clip-roberta-finetuned")

# Example usage
inputs = processor(text=["a photo of a cat"], images=["path_to_image.jpg"], return_tensors="pt", padding=True)
outputs = model(**inputs)
logits_per_image = outputs.logits_per_image

Training Details

Training Data

The model was fine-tuned on a custom dataset containing paired image-text data, focusing on diverse skin tones and various diseases. This dataset includes high-resolution images of dermatological conditions along with descriptive captions. The training data was carefully preprocessed to ensure high quality, with images resized, normalized, and tokenized text inputs.

Training Procedure

The fine-tuning process was conducted on a GPU-accelerated environment, leveraging the following setup and hyperparameters:

Preprocessing

  • Images: Resized to the appropriate dimensions, normalized according to the standard mean and standard deviation used in CLIP models.
  • Text: Tokenized using the roberta-base tokenizer.

Training Hyperparameters

  • Training regime: Mixed precision (fp16) training to optimize memory and computation.
  • Batch size: 32 for both training and evaluation.
  • Learning rate: 3e-5, with a cosine learning rate schedule with restarts.
  • Weight decay: 0.01 to prevent overfitting.
  • Warmup steps: 1000 steps to stabilize the learning process.
  • Number of epochs: 1000 to ensure thorough training.
  • Gradient accumulation steps: 4 to effectively simulate a larger batch size.
  • Evaluation strategy: Performed at the end of each epoch.
  • Logging strategy: Metrics logged at the end of each epoch.
  • Checkpointing: The best model was saved based on evaluation loss, with a maximum of 3 checkpoints retained.

Speeds, Sizes, Times

  • Training time: [More Information Needed]
  • Total epochs: 1000
  • Checkpoint size: [More Information Needed]

Compute Resources

The fine-tuning process was conducted on a system with the following specifications:

  • GPU: NVIDIA V100
  • Memory Usage: Monitored with Weights & Biases, optimized through gradient accumulation and mixed precision.
  • Training duration: [More Information Needed]

Evaluation

Testing Data, Factors & Metrics

Testing Data

<--- The evaluation was conducted on a held-out test set, which included images and captions not seen during training. The test set was designed to be diverse, representing various skin tones and dermatological conditions. ---->

Metrics

The primary evaluation metric was eval_loss. Additionally, accuracy and recall metrics were used to assess the model's ability to correctly match images with their corresponding textual descriptions.

Results

  • Best evaluation loss: [More Information Needed]
  • Top-1 accuracy: [More Information Needed]
  • Recall@k: [More Information Needed]

Summary

The model showed robust performance on the test set, particularly excelling in scenarios where both the image and text inputs were well-aligned with the training data. However, performance may vary with data significantly different from the training set.

Environmental Impact

Carbon emissions were monitored using Weights & Biases, with the following estimates:

  • Hardware Type: NVIDIA V100 GPU
  • Training Hours: [More Information Needed]
  • Cloud Provider: [More Information Needed]
  • Compute Region: [More Information Needed]
  • Carbon Emitted: [More Information Needed]

Technical Specifications

Model Architecture and Objective

The model integrates CLIP's vision encoder (clip-vit-base-patch32) with RoBERTa's language model (roberta-base) in a dual encoder setup. The objective was to fine-tune this model to improve its performance on image-text matching tasks, specifically within the medical domain, focusing on dermatological images.

Compute Infrastructure

Hardware

  • GPUs: NVIDIA V100
  • Memory: 16GB GPU memory, utilized efficiently with mixed precision training.

Software

  • Transformers version: [More Information Needed]
  • PyTorch version: [More Information Needed]

Citation

BibTeX:

@inproceedings{parthasarathy2024cliproberta,
  title={Fine-Tuned Vision-Text Dual Encoder Model for Image-Text Matching},
  author={Parthasarathy, Dhruv},
  year={2024},
  howpublished={\url{https://huggingface.co/your-username/clip-roberta-finetuned}},
}

Model Card Authors

This model card was prepared by Dhruv Parthasarathy.

Model Card Contact

For questions or issues with the model, please contact parthasarathy.d@northeastern.edu or linkedin.com/in/parthadhruv/.

Downloads last month
5
Safetensors
Model size
213M params
Tensor type
F32
·
Inference Examples
This model does not have enough activity to be deployed to Inference API (serverless) yet. Increase its social visibility and check back later, or deploy to Inference Endpoints (dedicated) instead.