ShareCaptioner Model Card

Model details

Model type: ShareCaptioner is an open-source captioner fine-tuned on GPT4-Vision-assisted ShareGPT4V detailed caption data with a resolution of 448x448. ShareCaptioner is based on the improved InternLM-Xcomposer-7B base model.

Model date: ShareCaptioner was trained in Nov 2023.

Paper or resources for more information: [Project] [Paper] [Code]

License

Llama 2 is licensed under the LLAMA 2 Community License, Copyright (c) Meta Platforms, Inc. All Rights Reserved.

Intended use

Primary intended uses: The primary use of ShareCaptioner is about producing high-quality image captions.

Primary intended users: The primary intended users of the model are researchers and hobbyists in computer vision, natural language processing, machine learning, and artificial intelligence.

Finetuning dataset

  • 100K GPT4-Vision-generated image-text pairs
Downloads last month
24
Inference Providers NEW
This model is not currently available via any of the supported Inference Providers.
The model cannot be deployed to the HF Inference API: The model authors have turned it off explicitly.