Edit model card

DenseConnector-v1.5-8B Model Card

Model details

Model type: DenseConnector is an open-source chatbot trained by fine-tuning LLaMA/Vicuna on GPT-generated multimodal instruction-following data. It is an auto-regressive language model, based on the transformer architecture.

Model date: DenseConnector-v1.5-8B was trained in 05/2024.

Paper or resources for more information: https://github.com/HJYao00/DenseConnector

Paper on Hugging Face: arxiv.org/abs/2405.13800

Training datase:t This model is trained on LLaVA-1.5 dataset.

Large Language Model: Llama-3-8B-Instruct

License

Llama 3 is licensed under the LLAMA 3 Community License, Copyright (c) Meta Platforms, Inc. All Rights Reserved.

Where to send questions or comments about the model: https://github.com/HJYao00/DenseConnector/issues

Intended use

Primary intended uses: The primary use of DenseConnector is research on large multimodal models and chatbots.

Primary intended users: The primary intended users of the model are researchers and hobbyists in computer vision, natural language processing, machine learning, and artificial intelligence.

Downloads last month
29
Safetensors
Model size
8.46B params
Tensor type
BF16
·
Inference Examples
This model does not have enough activity to be deployed to Inference API (serverless) yet. Increase its social visibility and check back later, or deploy to Inference Endpoints (dedicated) instead.

Spaces using HuanjinYao/DenseConnector-v1.5-8B 3

Collection including HuanjinYao/DenseConnector-v1.5-8B