|
--- |
|
license: llama3 |
|
library_name: transformers |
|
tags: |
|
- nsfw |
|
- not-for-all-audiences |
|
- llama-3 |
|
- text-generation-inference |
|
- moe |
|
- mergekit |
|
- merge |
|
--- |
|
|
|
# Llama-Salad-4x8B-V2 |
|
Changes in V2: |
|
- Swapped Tess-2.0-Llama-3-8B for Llama-3-8B-Synthia-v3.5 |
|
- Swapped L3-8B-Stheno-v3.1 for Llama-3-Soliloquy-8B-v2 |
|
- Removed Llama3-OpenBioLLM-8B and added opus-v1.2-llama-3-8b-instruct-run3.5-epoch2.5 |
|
|
|
V2 has improvements in all areas from V1; it's not a massive improvement, but I can confidently say it's a direct upgrade. Llama-3-8B-Synthia-v3.5 is better than Tess-2.0-Llama-3-8B in every way; Llama-3-Soliloquy-8B-v2 is more intelligent than L3-8B-Stheno-v3.1 and has less bias towards NSFW content; and the inclusion of opus-v1.2-llama-3-8b-instruct-run3.5-epoch2.5 has greatly improved its storytelling and narration abilities. |
|
|
|
I really like the model selection in this one, so I don't know how much more I can improve if I make another 4x8B merge. If I were to make a V3, swapping Meta-Llama-3-8B-Instruct would likely be the only change. I will try my hand at making an 8x8B merge in the future, but I still need to find some models to fill the gaps; making sure there's no routing conflicts between 8 different models at once will be the biggest challenge. |
|
|
|
# Quantization Formats |
|
**GGUF** |
|
- Static: |
|
- https://huggingface.co/mradermacher/Llama-Salad-4x8B-V2-GGUF |
|
- Imatrix: |
|
- https://huggingface.co/mradermacher/Llama-Salad-4x8B-V2-i1-GGUF |
|
|
|
# Details |
|
- **License**: [llama3](https://llama.meta.com/llama3/license/) |
|
- **Instruct Format**: [llama-3](https://llama.meta.com/docs/model-cards-and-prompt-formats/meta-llama-3/) |
|
- **Context Size**: 8K |
|
|
|
## Models Used |
|
- [Meta-Llama-3-8B-Instruct](https://huggingface.co/NousResearch/Meta-Llama-3-8B-Instruct) |
|
- [Llama-3-8B-Synthia-v3.5](https://huggingface.co/migtissera/Llama-3-8B-Synthia-v3.5) |
|
- [Llama-3-Soliloquy-8B-v2](https://huggingface.co/openlynn/Llama-3-Soliloquy-8B-v2) |
|
- [opus-v1.2-llama-3-8b-instruct-run3.5-epoch2.5](https://huggingface.co/dreamgen-preview/opus-v1.2-llama-3-8b-instruct-run3.5-epoch2.5) |
|
|
|
## Merge Config |
|
```yaml |
|
base_model: NousResearch/Meta-Llama-3-8B-Instruct |
|
gate_mode: hidden |
|
dtype: bfloat16 |
|
experts_per_token: 2 |
|
experts: |
|
- source_model: NousResearch/Meta-Llama-3-8B-Instruct |
|
positive_prompts: |
|
- "summarize" |
|
- "paraphrase" |
|
- "explain" |
|
- "define" |
|
- "translate" |
|
- "multilingual" |
|
- "chat" |
|
- "conversation" |
|
- source_model: migtissera/Llama-3-8B-Synthia-v3.5 |
|
positive_prompts: |
|
- "programming language" |
|
- "JavaScript" |
|
- "Python programming language" |
|
- "Rust programming language" |
|
- "CSS markup styling language" |
|
- "math" |
|
- "code" |
|
- "step-by-step" |
|
- "logical reasoning" |
|
- source_model: openlynn/Llama-3-Soliloquy-8B-v2 |
|
positive_prompts: |
|
- "roleplay" |
|
- "erotic roleplay" |
|
- "characters" |
|
- "scene" |
|
- "opinion" |
|
- source_model: dreamgen-preview/opus-v1.2-llama-3-8b-instruct-run3.5-epoch2.5 |
|
positive_prompts: |
|
- "creative writing" |
|
- "storytelling" |
|
- "narration" |
|
- "narrative setting" |
|
- "narrative plot" |
|
- "narrative exposition" |
|
- "narrative theme" |
|
- "narrative climax" |
|
``` |