|
--- |
|
license: llama2 |
|
model_name: Vicuna 7B v1.5 |
|
base_model: lmsys/vicuna-7b-v1.5 |
|
inference: false |
|
model_creator: lmsys |
|
model_type: llama |
|
prompt_template: > |
|
A chat between a curious user and an artificial intelligence assistant. The |
|
assistant gives helpful, detailed, and polite answers to the user's questions. |
|
USER: {prompt} ASSISTANT: |
|
prepared_by: TheMarmot |
|
--- |
|
## Model Details |
|
|
|
### Model Description |
|
|
|
This is an unquantized version of the Vicuna 7B v1.5 from lmsys in the gguf format which is compatible with llama-cpp. |
|
|
|
- **Developed by:** [Joseph Bejjani] |
|
|
|
|