caesar-one's picture
Update README.md
1ec5c98 verified
metadata
license: apache-2.0
base_model: mistralai/Mistral-7B-Instruct-v0.2
tags:
  - trl
  - sft
  - generated_from_trainer
datasets:
  - generator
model-index:
  - name: DanteLLM_instruct_7b-v0.2-boosted
    results: []
library_name: peft

DanteLLM

DanteLLM is a Large Language Model developed in Sapienza lab. In October 2023 we submitted a paper called DanteLLM: Let's Push Italian LLM Research Forward! 🤌 🇮🇹

That paper got accepted with the scores 5, 4, 4 out of 5

How to run the model (Ollama)

This repo contains the model in GGUF format. You can run DanteLLM on Ollama following these steps:

Make sure you have Ollama correctly installed and ready to use.

Then, you can download DanteLLM's weights using:

huggingface-cli download rstless-research/DanteLLM-7B-Instruct-Italian-v0.1-GGUF dantellm-merged-hf.q8_0.gguf Modelfile --local-dir . --local-dir-use-symlinks False

Load the model using:

ollama create dante -f Modelfile

Finally, to run the model, use:

ollama run dante

Authors

  • Andrea Bacciu* (work done prior joining Amazon)
  • Cesare Campagnano*
  • Giovanni Trappolini
  • Prof. Fabrizio Silvestri

* Equal contribution