Edit model card

Triangle104/Llama3.2-3B-ShiningValiant2-Q8_0-GGUF

This model was converted to GGUF format from ValiantLabs/Llama3.2-3B-ShiningValiant2 using llama.cpp via the ggml.ai's GGUF-my-repo space. Refer to the original model card for more details on the model.


Model details:

Shining Valiant 2 is a chat model built on Llama 3.2 3b, finetuned on our data for friendship, insight, knowledge and enthusiasm.

Finetuned on meta-llama/Llama-3.2-3B-Instruct for best available general performance
Trained on a variety of high quality data; focused on science, engineering, technical knowledge, and structured reasoning
Also available for Llama 3.1 70b and Llama 3.1 8b!

Version

This is the 2024-09-27 release of Shining Valiant 2 for Llama 3.2 3b.

We've improved and open-sourced our new baseline science-instruct dataset. This release features improvements in physics, chemistry, biology, and computer science.

Future upgrades will continue to expand Shining Valiant's technical knowledge base.

Help us and recommend Shining Valiant 2 to your friends! Prompting Guide

Shining Valiant 2 uses the Llama 3.2 Instruct prompt format. The example script below can be used as a starting point for general chat:

import transformers import torch

model_id = "ValiantLabs/Llama3.2-3B-ShiningValiant2"

pipeline = transformers.pipeline( "text-generation", model=model_id, model_kwargs={"torch_dtype": torch.bfloat16}, device_map="auto", )

messages = [ {"role": "system", "content": "You are an AI assistant."}, {"role": "user", "content": "Describe the use of chiral auxiliaries in organic synthesis."} ]

outputs = pipeline( messages, max_new_tokens=2048, )

print(outputs[0]["generated_text"][-1])

The Model

Shining Valiant 2 is built on top of Llama 3.2 3b Instruct.

The current version of Shining Valiant 2 is trained on technical knowledge using sequelbox/Celestia, complex reasoning using sequelbox/Spurline, and general chat capability using sequelbox/Supernova.

We're super excited that Shining Valiant's dataset has been fully open-sourced! She's friendly, enthusiastic, insightful, knowledgeable, and loves to learn! Magical.

Shining Valiant 2 is created by Valiant Labs.

Check out our HuggingFace page for our open-source Build Tools models, including the newest version of code-specialist Enigma!

Follow us on X for updates on our models!

We care about open source. For everyone to use.

We encourage others to finetune further from our models.


Use with llama.cpp

Install llama.cpp through brew (works on Mac and Linux)

brew install llama.cpp

Invoke the llama.cpp server or the CLI.

CLI:

llama-cli --hf-repo Triangle104/Llama3.2-3B-ShiningValiant2-Q8_0-GGUF --hf-file llama3.2-3b-shiningvaliant2-q8_0.gguf -p "The meaning to life and the universe is"

Server:

llama-server --hf-repo Triangle104/Llama3.2-3B-ShiningValiant2-Q8_0-GGUF --hf-file llama3.2-3b-shiningvaliant2-q8_0.gguf -c 2048

Note: You can also use this checkpoint directly through the usage steps listed in the Llama.cpp repo as well.

Step 1: Clone llama.cpp from GitHub.

git clone https://github.com/ggerganov/llama.cpp

Step 2: Move into the llama.cpp folder and build it with LLAMA_CURL=1 flag along with other hardware-specific flags (for ex: LLAMA_CUDA=1 for Nvidia GPUs on Linux).

cd llama.cpp && LLAMA_CURL=1 make

Step 3: Run inference through the main binary.

./llama-cli --hf-repo Triangle104/Llama3.2-3B-ShiningValiant2-Q8_0-GGUF --hf-file llama3.2-3b-shiningvaliant2-q8_0.gguf -p "The meaning to life and the universe is"

or

./llama-server --hf-repo Triangle104/Llama3.2-3B-ShiningValiant2-Q8_0-GGUF --hf-file llama3.2-3b-shiningvaliant2-q8_0.gguf -c 2048
Downloads last month
24
GGUF
Model size
3.21B params
Architecture
llama

8-bit

Inference Examples
Unable to determine this model's library. Check the docs .

Model tree for Triangle104/Llama3.2-3B-ShiningValiant2-Q8_0-GGUF

Datasets used to train Triangle104/Llama3.2-3B-ShiningValiant2-Q8_0-GGUF

Collection including Triangle104/Llama3.2-3B-ShiningValiant2-Q8_0-GGUF

Evaluation results