Edit model card

TinyLlama Dolly 15

The TinyLlama Dolly 15k is a specialized large language model finely tuned on the Databricks Dolly 15k dataset. This dataset, composed of 15,000 high-quality, human-curated prompts and responses, helps the model excel in understanding and generating contextually relevant and coherent text. By leveraging this diverse and comprehensive dataset, TinyLlama Dolly 15k enhances its ability to engage in nuanced conversations and provide accurate, context-aware responses across a range of topics. The result is a powerful tool for applications requiring advanced natural language understanding and generation.

How to use

Below is a snippet that can be used to test the model.

from transformers import pipeline

pipe = pipeline("text-generation", model="Bhooyas/tinyllama-dolly-15k", device_map="auto")

messages = [
    {
        "role": "user",
        "context": "The TinyLlama Dolly 15k is a specialized large language model finely tuned on the Databricks Dolly 15k dataset. This dataset, composed of 15,000 high-quality, human-curated prompts and responses, helps the model excel in understanding and generating contextually relevant and coherent text. By leveraging this diverse and comprehensive dataset, TinyLlama Dolly 15k enhances its ability to engage in nuanced conversations and provide accurate, context-aware responses across a range of topics. The result is a powerful tool for applications requiring advanced natural language understanding and generation.",
        "instruction": "What is TinyLlama Dolly 15k?"
    }
]

prompt = pipe.tokenizer.apply_chat_template(messages, tokenize=False, add_generation_prompt=True)
outputs = pipe(prompt, max_new_tokens=256)
print(outputs[0]["generated_text"])

Downloads last month
8
Safetensors
Model size
1.1B params
Tensor type
F32
·
Inference Examples
Unable to determine this model's library. Check the docs .

Model tree for Bhooyas/tinyllama-dolly-15k

Finetuned
(152)
this model

Dataset used to train Bhooyas/tinyllama-dolly-15k