metadata
license: apache-2.0
base_model: mistralai/Mistral-7B-Instruct-v0.2
tags:
- trl
- sft
- generated_from_trainer
datasets:
- generator
model-index:
- name: DanteLLM_instruct_7b-v0.2-boosted
results: []
library_name: peft
DanteLLM
DanteLLM is a Large Language Model developed in Sapienza lab. In October 2023 we submitted a paper called DanteLLM: Let's Push Italian LLM Research Forward! 🤌 🇮🇹
That paper got accepted with the scores 5, 4, 4 out of 5
How to run the model (Ollama)
This repo contains the model in GGUF format. You can run DanteLLM on Ollama following these steps:
Make sure you have Ollama correctly installed and ready to use.
Then, you can download DanteLLM's weights using:
huggingface-cli download rstless-research/DanteLLM-7B-Instruct-Italian-v0.1-GGUF dantellm-merged-hf.q8_0.gguf Modelfile --local-dir . --local-dir-use-symlinks False
Load the model using:
ollama create dante -f Modelfile
Finally, to run the model, use:
ollama run dante
Authors
- Andrea Bacciu* (work done prior joining Amazon)
- Cesare Campagnano*
- Giovanni Trappolini
- Prof. Fabrizio Silvestri
* Equal contribution