TinyLlama Dolly 15
The TinyLlama Dolly 15k is a specialized large language model finely tuned on the Databricks Dolly 15k dataset. This dataset, composed of 15,000 high-quality, human-curated prompts and responses, helps the model excel in understanding and generating contextually relevant and coherent text. By leveraging this diverse and comprehensive dataset, TinyLlama Dolly 15k enhances its ability to engage in nuanced conversations and provide accurate, context-aware responses across a range of topics. The result is a powerful tool for applications requiring advanced natural language understanding and generation.
How to use
Below is a snippet that can be used to test the model.
from transformers import pipeline
pipe = pipeline("text-generation", model="Bhooyas/tinyllama-dolly-15k", device_map="auto")
messages = [
{
"role": "user",
"context": "The TinyLlama Dolly 15k is a specialized large language model finely tuned on the Databricks Dolly 15k dataset. This dataset, composed of 15,000 high-quality, human-curated prompts and responses, helps the model excel in understanding and generating contextually relevant and coherent text. By leveraging this diverse and comprehensive dataset, TinyLlama Dolly 15k enhances its ability to engage in nuanced conversations and provide accurate, context-aware responses across a range of topics. The result is a powerful tool for applications requiring advanced natural language understanding and generation.",
"instruction": "What is TinyLlama Dolly 15k?"
}
]
prompt = pipe.tokenizer.apply_chat_template(messages, tokenize=False, add_generation_prompt=True)
outputs = pipe(prompt, max_new_tokens=256)
print(outputs[0]["generated_text"])
- Downloads last month
- 8
Model tree for Bhooyas/tinyllama-dolly-15k
Base model
TinyLlama/TinyLlama-1.1B-Chat-v1.0