Edit model card

LLaVA_X_KoLlama2-7B-0.1v

img

KoT-platypus2 X LLaVA

This model is a large multimodal model (LMM) that combines the LLM(KoT-platypus2-7B) with visual encoder of CLIP(ViT-14), trained on Korean visual-instruction dataset using QLoRA.

Model Details

  • Model Developers: Nagase_Kotono
  • Base Model: kyujinpy/KoT-platypus2-7B
  • Model Architecture: LLaVA_X_KoLlama2-7B is an open-source chatbot trained by fine-tuning Llama2 on GPT-generated multimodal instruction-following data. It is an auto-regressive language model, based on the Llama2 transformer architecture.
  • Training Dataset: KoLLaVA-CC3M-Pretrain-595K, KoLLaVA-Instruct-150k

Hardware & Software

Pretrain

Finetune

ValueError

LLaVA_X_KoLlama2-7B is a base model of kyujinpy/KoT-platypus2-7B. The model is based on beomi/llama-2-ko-7b.
Since Llama-2-Ko uses FastTokenizer provided by HF tokenizers NOT sentencepiece package, it is required to use use_fast=True option when initialize tokenizer.

Downloads last month
15
Inference Examples
This model does not have enough activity to be deployed to Inference API (serverless) yet. Increase its social visibility and check back later, or deploy to Inference Endpoints (dedicated) instead.