File size: 2,278 Bytes
b4172ab
f238a00
b4172ab
bab5a9f
f238a00
 
 
267e97e
f238a00
bab5a9f
267e97e
 
8fb9ac0
267e97e
8fb9ac0
267e97e
8fb9ac0
 
 
 
267e97e
8fb9ac0
267e97e
8fb9ac0
 
 
 
267e97e
8fb9ac0
f238a00
267e97e
 
 
8fb9ac0
 
 
 
 
 
 
 
 
 
 
 
 
 
267e97e
8fb9ac0
 
267e97e
f238a00
 
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
---
inference: false
---

**NOTE: New version available**  
Please check out a newer version of the weights [here](https://huggingface.co/lmsys/vicuna-7b-v1.3).  
If you still want to use this old version, please see the compatibility and difference between different versions [here](https://github.com/lm-sys/FastChat/blob/main/docs/vicuna_weights_version.md).

<br>

# Vicuna Model Card

## Model Details

Vicuna is a chat assistant trained by fine-tuning LLaMA on user-shared conversations collected from ShareGPT.

- **Developed by:** [LMSYS](https://lmsys.org/)
- **Model type:** An auto-regressive language model based on the transformer architecture.
- **License:** Non-commercial license
- **Finetuned from model:** [LLaMA](https://arxiv.org/abs/2302.13971).

### Model Sources

- **Repository:** https://github.com/lm-sys/FastChat
- **Blog:** https://lmsys.org/blog/2023-03-30-vicuna/
- **Paper:** https://arxiv.org/abs/2306.05685
- **Demo:** https://chat.lmsys.org/

## Uses

The primary use of Vicuna is research on large language models and chatbots.
The primary intended users of the model are researchers and hobbyists in natural language processing, machine learning, and artificial intelligence.

## How to Get Started with the Model

Command line interface: https://github.com/lm-sys/FastChat#vicuna-weights.  
APIs (OpenAI API, Huggingface API): https://github.com/lm-sys/FastChat/tree/main#api.  

## Training Details

Vicuna v1.1 is fine-tuned from LLaMA with supervised instruction fine-tuning.
The training data is around 70K conversations collected from ShareGPT.com.
See more details in the "Training Details of Vicuna Models" section in the appendix of this [paper](https://arxiv.org/pdf/2306.05685.pdf).

## Evaluation

Vicuna is evaluated with standard benchmarks, human preference, and LLM-as-a-judge. See more details in this [paper](https://arxiv.org/pdf/2306.05685.pdf) and [leaderboard](https://huggingface.co/spaces/lmsys/chatbot-arena-leaderboard).

## Difference between different versions of Vicuna
See [vicuna_weights_version.md](https://github.com/lm-sys/FastChat/blob/main/docs/vicuna_weights_version.md)

## Acknowledgement
Special thanks to [@TheBloke](https://huggingface.co/TheBloke) for hosting this merged version of weights earlier.