Edit model card

GenZ 13B v2 GGML

The instruction finetuned model with 4K input length. The model is finetuned on top of pretrained LLaMa2

Inference

import ctransformers
from ctransformers import AutoModelForCausalLM


model = AutoModelForCausalLM.from_pretrained('budecosystem/genz-13b-v2-ggml', model_type="llama")

prompt = """A chat between a curious user and an artificial intelligence assistant. The assistant gives helpful, detailed, and polite answers to the user's questions.
USER: who are you? ASSISTANT: """
print(model(prompt))

Support LM Studio for Mac & Windows users

Use following prompt template

A chat between a curious user and an artificial intelligence assistant. The assistant gives helpful, detailed, and polite answers to the user's questions. USER: Hi, how are you? ASSISTANT: 

Check the GitHub for the code -> GenZ

Downloads last month
12
Inference Examples
This model does not have enough activity to be deployed to Inference API (serverless) yet. Increase its social visibility and check back later, or deploy to Inference Endpoints (dedicated) instead.