--- base_model: [] library_name: transformers tags: - mergekit - merge --- # miquplus-xwin-70b This is a merge of pre-trained language models created using [mergekit](https://github.com/cg123/mergekit). ## Merge Details ### Merge Method This model was merged using the [linear](https://arxiv.org/abs/2203.05482) merge method using /home/juk/LLMs/models/huggingface/miqu-1-70b-sf as a base. ### Models Merged The following models were included in the merge: * miqu-models/_miquplus-xwin-70b ### Configuration The following YAML configuration was used to produce this model: ```yaml base_model: model: path: /home/juk/LLMs/models/huggingface/miqu-1-70b-sf dtype: float16 merge_method: linear slices: - sources: - layer_range: [0, 80] model: model: path: /home/juk/LLMs/models/huggingface/miqu-1-70b-sf parameters: weight: - filter: v_proj value: [1.0, 1.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 1.0, 1.0] - filter: o_proj value: [1.0, 1.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 1.0, 1.0] - filter: up_proj value: [1.0, 1.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 1.0, 1.0] - filter: gate_proj value: [1.0, 1.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 1.0, 1.0] - filter: down_proj value: [1.0, 1.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 1.0, 1.0] - value: 1.0 - layer_range: [0, 80] model: model: path: miqu-models/_miquplus-xwin-70b parameters: weight: - filter: v_proj value: [0.0, 0.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 0.0, 0.0] - filter: o_proj value: [0.0, 0.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 0.0, 0.0] - filter: up_proj value: [0.0, 0.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 0.0, 0.0] - filter: gate_proj value: [0.0, 0.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 0.0, 0.0] - filter: down_proj value: [0.0, 0.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 0.0, 0.0] - value: 0.0 tokenizer_source: base ```