pere's picture
Upload README.md with huggingface_hub
0d24f20 verified
|
raw
history blame
2.99 kB
metadata
base_model: NB-Llama-3.2-1B-Instruct
language:
  - 'no'
  - nb
  - nn
  - en
  - sv
  - da
tags:
  - llama-cpp
  - gguf
  - quantization
  - norwegian
  - bokmål
  - nynorsk
  - swedish
  - danish
  - multilingual
  - text-generation
pipeline_tag: text-generation
license: llama3.2

NB-Llama-3.2-1B-Instruct-Q4_K_M-GGUF

This model is a quantized version of the original NB-Llama-3.2-1B-Instruct, converted into the GGUF format using llama.cpp. Quantization significantly reduces the model's memory footprint, enabling efficient inference on a wide range of hardware, including personal devices, without compromising too much quality. These quantized models are mainly provided so that people can test out the models with moderate hardware. If you want to benchmark the models or further finetune the models, we strongly recommend the non-quantized versions.

What is llama.cpp?

llama.cpp is a versatile tool for running large language models optimized for efficiency. It supports multiple quantization formats (e.g., GGML and GGUF) and provides inference capabilities on diverse hardware, including CPUs, GPUs, and mobile devices. The GGUF format is the latest evolution, designed to enhance compatibility and performance.

Benefits of This Model

  • High Performance: Achieves similar quality to the original model while using significantly less memory.
  • Hardware Compatibility: Optimized for running on a variety of hardware, including low-resource systems.
  • Ease of Use: Seamlessly integrates with llama.cpp for fast and efficient inference.

Installation

Install llama.cpp using Homebrew (works on Mac and Linux):

brew install llama.cpp

Usage Instructions

Using with llama.cpp

To use this quantized model with llama.cpp, follow the steps below:

CLI:

llama-cli --hf-repo NbAiLab/nb-llama-3.2-1B-Instruct-Q4_K_M-GGUF --hf-file nb-llama-3.2-1b-instruct-q4_k_m.gguf -p "Your prompt here"

Server:

llama-server --hf-repo NbAiLab/nb-llama-3.2-1B-Instruct-Q4_K_M-GGUF --hf-file nb-llama-3.2-1b-instruct-q4_k_m.gguf -c 2048

For more information, refer to the llama.cpp repository.

Additional Resources

Citing & Authors

The model was trained and documentation written by Per Egil Kummervold

Funding and Acknowledgement

Training this model was supported by Google’s TPU Research Cloud (TRC), which generously supplied us with Cloud TPUs essential for our computational needs..