base_model: []
library_name: transformers
tags:
- mergekit
- merge
merge
This is a merge of pre-trained language models created using mergekit.
Thank you to MatchaSpeed for picking out the models to go into this merge. You're awesome! 🤗
The models in this merge were chosen for their attention to detail and/or their creativity.
Based on each model's individual strength, weights were applied to prioritize the model's intelligence (self-attention weights) or style/creativity (multi-layer perceptron weights).
GGUF versions available courtesy of BackyardAI https://huggingface.co/backyardai/BlackMagic-7B-GGUF
Usage
Recommended prompt format is Alpaca, as this is most widely accepted by the component models.
Merge Details
Merge Method
This model was merged using the task arithmetic merge method using Mistral-7B-v0.1 as a base.
Models Merged
The following models were included in the merge:
- fearlessdots/WizardLM-2-7B-abliterated
- Gryphe/MythoMist-7b
- Sao10K/Frostwind-v2.1-m7
- senseable/Westlake-7b-v2
- maywell/PiVoT-0.1-Evil-a
- Undi95/Toppy-M-7B
Configuration
The following YAML configuration was used to produce this model:
models:
- model: fearlessdots/WizardLM-2-7B-abliterated
parameters:
weight: 1.0
- model: Undi95/Toppy-M-7B
parameters:
weight:
- filter: self_attn
value: 0.8
- value: 0.5
- model: senseable/Westlake-7b-v2
parameters:
weight:
- filter: self_attn
value: 0.6
- value: 0.4
- model: maywell/PiVoT-0.1-Evil-a
parameters:
weight:
- filter: mlp
value: 0.2
- value: 0.0
- model: Sao10K/Frostwind-v2.1-m7
parameters:
weight:
- filter: self_attn
value: 0.2
- filter: mlp
value: 0.8
- value: 0.5
- model: Gryphe/MythoMist-7b
parameters:
weight:
- filter: mlp
value: 0.6
- value: 0.0
base_model: TeeZee/Mistral-7B-v0.1-fp32
merge_method: task_arithmetic
dtype: float32