|
--- |
|
license: apache-2.0 |
|
base_model: mistralai/Mistral-7B-Instruct-v0.2 |
|
tags: |
|
- trl |
|
- sft |
|
- generated_from_trainer |
|
datasets: |
|
- generator |
|
model-index: |
|
- name: DanteLLM_instruct_7b-v0.2-boosted |
|
results: [] |
|
library_name: peft |
|
--- |
|
|
|
<!-- This model card has been generated automatically according to the information the Trainer had access to. You |
|
should probably proofread and complete it, then remove this comment. --> |
|
## DanteLLM |
|
|
|
DanteLLM is a Large Language Model developed in Sapienza lab. |
|
In October 2023 we submitted a paper called DanteLLM: Let's Push Italian LLM Research Forward! |
|
๐ค ๐ฎ๐น |
|
|
|
|
|
That paper got accepted with the scores 5, 4, 4 out of 5 |
|
## How to run the model (Ollama) |
|
This repo contains the model in GGUF format. You can run DanteLLM on Ollama following these steps: |
|
|
|
Make sure you have Ollama correctly installed and ready to use. |
|
|
|
Then, you can download DanteLLM's weights using: |
|
``` |
|
huggingface-cli download rstless-research/DanteLLM-7B-Instruct-Italian-v0.1-GGUF dantellm-merged-hf.q8_0.gguf Modelfile --local-dir . --local-dir-use-symlinks False |
|
``` |
|
Load the model using: |
|
``` |
|
ollama create dante -f Modelfile |
|
``` |
|
Finally, to run the model, use: |
|
``` |
|
ollama run dante |
|
``` |
|
|
|
# Authors |
|
- Andrea Bacciu* (work done prior joining Amazon) |
|
- Cesare Campagnano* |
|
- Giovanni Trappolini |
|
- Prof. Fabrizio Silvestri |
|
|
|
\* Equal contribution |