--- license: apache-2.0 base_model: mistralai/Mistral-7B-Instruct-v0.2 tags: - trl - sft - generated_from_trainer datasets: - generator model-index: - name: DanteLLM_instruct_7b-v0.2-boosted results: [] library_name: peft --- ## DanteLLM DanteLLM is a Large Language Model developed in Sapienza lab. In October 2023 we submitted a paper called DanteLLM: Let's Push Italian LLM Research Forward! 🤌 🇮🇹 That paper got accepted with the scores 5, 4, 4 out of 5 ## How to run the model (Ollama) This repo contains the model in GGUF format. You can run DanteLLM on Ollama following these steps: Make sure you have Ollama correctly installed and ready to use. Then, you can download DanteLLM's weights using: ``` huggingface-cli download rstless-research/DanteLLM-7B-Instruct-Italian-v0.1-GGUF dantellm-merged-hf.q8_0.gguf Modelfile --local-dir . --local-dir-use-symlinks False ``` Load the model using: ``` ollama create dante -f Modelfile ``` Finally, to run the model, use: ``` ollama run dante ``` # Authors - Andrea Bacciu* (work done prior joining Amazon) - Cesare Campagnano* - Giovanni Trappolini - Prof. Fabrizio Silvestri \* Equal contribution