caesar-one's picture
Update README.md
1ec5c98 verified
---
license: apache-2.0
base_model: mistralai/Mistral-7B-Instruct-v0.2
tags:
- trl
- sft
- generated_from_trainer
datasets:
- generator
model-index:
- name: DanteLLM_instruct_7b-v0.2-boosted
results: []
library_name: peft
---
<!-- This model card has been generated automatically according to the information the Trainer had access to. You
should probably proofread and complete it, then remove this comment. -->
## DanteLLM
DanteLLM is a Large Language Model developed in Sapienza lab.
In October 2023 we submitted a paper called DanteLLM: Let's Push Italian LLM Research Forward!
๐ŸคŒ ๐Ÿ‡ฎ๐Ÿ‡น
That paper got accepted with the scores 5, 4, 4 out of 5
## How to run the model (Ollama)
This repo contains the model in GGUF format. You can run DanteLLM on Ollama following these steps:
Make sure you have Ollama correctly installed and ready to use.
Then, you can download DanteLLM's weights using:
```
huggingface-cli download rstless-research/DanteLLM-7B-Instruct-Italian-v0.1-GGUF dantellm-merged-hf.q8_0.gguf Modelfile --local-dir . --local-dir-use-symlinks False
```
Load the model using:
```
ollama create dante -f Modelfile
```
Finally, to run the model, use:
```
ollama run dante
```
# Authors
- Andrea Bacciu* (work done prior joining Amazon)
- Cesare Campagnano*
- Giovanni Trappolini
- Prof. Fabrizio Silvestri
\* Equal contribution