# LoRA Adapter Model This is a LoRA adapter model fine-tuned on llava-hf/llava-1.5-7b-hf. ## Model Details - Base Model: llava-hf/llava-1.5-7b-hf - Training Parameters: - Learning Rate: 1e-4 - Batch Size: 16 - Training Steps: 58 ## Usage ```python from transformers import LlavaForConditionalGeneration, AutoProcessor from peft import PeftModel import torch # Load base model base_model = LlavaForConditionalGeneration.from_pretrained( "llava-hf/llava-1.5-7b-hf", revision='a272c74', torch_dtype=torch.float16, device_map="auto" ) tokenizer = AutoProcessor.from_pretrained("llava-hf/llava-1.5-7b-hf", revision='a272c74') # Load LoRA adapter model = PeftModel.from_pretrained( base_model, "Dipto084/RepLLaVA4", torch_dtype=torch.float16, device_map="auto" ) ```