medical-gen-small / README.md
Ra-Is's picture
updated
0c541b2 verified
metadata
language: en
license: apache-2.0
pipeline_tag: text-generation
base_model: t5-small
library_name: transformers
widget:
  - text: A 35-year-old female presents with a 2-week history of persistent cough...

Medical Generation Model

Overview

This repository contains a fine-tuned T5 model designed to generate medical diagnoses and treatment recommendations. The model was trained on clinical scenarios to provide accurate and contextually relevant medical outputs based on input prompts.

Model Details

  • Model Type: T5
  • Tokenizer: T5 tokenizer
  • Training Data: Clinical scenarios and medical texts

Installation

To use this model, install the required libraries with pip:

pip install transformers
pip install tensorflow

# Load the fine-tuned model and tokenizer
from transformers import T5Tokenizer, TFT5ForConditionalGeneration

model_id = "Ra-Is/medical-gen-small"
model = TFT5ForConditionalGeneration.from_pretrained(model_id)
tokenizer = T5Tokenizer.from_pretrained(model_id)

# Prepare a sample input prompt
input_prompt = ("A 35-year-old female presents with a 2-week history of "
                "persistent cough, shortness of breath, and fatigue. She has "
                "a history of asthma and has recently been exposed to a sick "
                "family member with a respiratory infection. Chest X-ray shows "
                "bilateral infiltrates. What is the likely diagnosis, and what "
                "should be the treatment?")

# Tokenize the input
input_ids = tokenizer(input_prompt, return_tensors="tf").input_ids

# Generate the output (diagnosis)
outputs = model.generate(
        input_ids,
        max_length=512,
        num_beams=5,
        temperature=1,
        top_k=50,
        top_p=0.9,
        do_sample=True,  # Enable sampling
        early_stopping=True
    )

# Decode and print the output
generated_text = tokenizer.decode(outputs[0], skip_special_tokens=True)
print(generated_text)