Edit model card

Uploaded model

  • Developed by: vaiditya
  • License: apache-2.0
  • Finetuned from model : meta/meta-llama-3.1-8b-bnb-4bit

How to Run the Model (Easiest Version)

  • Ollama You can run this model direct with ollama
  • Step1 Download ollama for your operating system from https://ollama.com/
  • Step2 Run the following command to run the model it will automatically download the model and execute it
  • Command
  • 4bitModel ollama run hf.co/vaiditya/codemelon-quantized-0.1-8b:Q4_K_M
  • 5bitModel ollama run hf.co/vaiditya/codemelon-quantized-0.1-8b:Q5_K_M
  • 8bitModel ollama run hf.co/vaiditya/codemelon-quantized-0.1-8b:Q8_0
  • 16bitModel ollama run hf.co/vaiditya/codemelon-quantized-0.1-8b:F16
Downloads last month
88
GGUF
Model size
8.03B params
Architecture
llama

4-bit

5-bit

8-bit

16-bit

Inference API
Unable to determine this model’s pipeline type. Check the docs .