|
--- |
|
license: apache-2.0 |
|
datasets: |
|
- ecastera/wiki_fisica |
|
- ecastera/filosofia-es |
|
- bertin-project/alpaca-spanish |
|
language: |
|
- es |
|
- en |
|
tags: |
|
- mistral |
|
- spanish |
|
- español |
|
- lora |
|
- int4 |
|
- multilingual |
|
--- |
|
|
|
# ecastera-eva-westlake-7b-spanish |
|
|
|
Mistral 7b-based model fine-tuned in Spanish to add high quality Spanish text generation. |
|
|
|
* Exported in GGUF format, INT4 quantization |
|
* Refined version of my previous models, with new training data and methodology. This should produce more natural reponses in Spanish. |
|
* Base model Mistral-7b |
|
* Based on the excelent job of senseable/WestLake-7B-v2 and Eric Hartford's cognitivecomputations/WestLake-7B-v2-laser |
|
* Fine-tuned in Spanish with a collection of poetry, books, wikipedia articles, phylosophy texts and alpaca-es datasets. |
|
* Trained using Lora and PEFT and INT8 quantization on 2 GPUs for several days. |
|
|
|
## Usage: |
|
|
|
Use in llamacpp or other framework that supports GGUF format. |
|
|
|
|