--- license: llama2 model_name: Vicuna 7B v1.5 base_model: lmsys/vicuna-7b-v1.5 inference: false model_creator: lmsys model_type: llama prompt_template: > A chat between a curious user and an artificial intelligence assistant. The assistant gives helpful, detailed, and polite answers to the user's questions. USER: {prompt} ASSISTANT: prepared_by: TheMarmot --- ## Model Details ### Model Description This is an unquantized version of the Vicuna 7B v1.5 from lmsys in the gguf format which is compatible with llama-cpp. - **Developed by:** [Joseph Bejjani]