Edit model card
[fp16 - gguf - exl2 : 2.55bpw]

Description

This repo contains fp16 files of Dawn-70B, a merge I have done with the new layer shuffle method from mergekit.

UtopiaXL was a huge success for me, I really liked it, so I took the same path to do this 70B: A good base, some psychologic data, some medical data, a little bit of this, of that, and LimaRP at the end as always.

NOTE: This repo contain the file measurement.json needed to do your own exl2 quant (I use wikitext).

Models and loras used

The sauce

!mergekit-layershuffle ./Dawn-v2-70B \
  --model Sao10K/Euryale-1.3-L2-70B --weight 0.3 \
  --model Xwin-LM/Xwin-LM-70B-V0.1 --weight 0.2 \
  --model ehartford/Samantha-1.11-70b --weight 0.1 \
  --model NousResearch/Nous-Hermes-Llama2-70b --weight 0.05 \
  --model augtoma/qCammel-70-x --weight 0.05 \
  --model jondurbin/airoboros-l2-c70b-3.1.2 --weight 0.2 \
  --model fangloveskari/ORCA_LLaMA_70B_QLoRA --weight 0.1 \
  --write-yaml Dawn-v2-70B.yaml

=========================

merge_method: passthrough
slices:
- sources:
  - layer_range:
    - 0
    - 1
    model: fangloveskari/ORCA_LLaMA_70B_QLoRA
- sources:
  - layer_range:
    - 1
    - 2
    model: jondurbin/airoboros-l2-c70b-3.1.2
- sources:
  - layer_range:
    - 2
    - 3
    model: Sao10K/Euryale-1.3-L2-70B
- sources:
  - layer_range:
    - 3
    - 4
    model: jondurbin/airoboros-l2-c70b-3.1.2
- sources:
  - layer_range:
    - 4
    - 5
    model: fangloveskari/ORCA_LLaMA_70B_QLoRA
- sources:
  - layer_range:
    - 5
    - 6
    model: ehartford/Samantha-1.11-70b
- sources:
  - layer_range:
    - 6
    - 8
    model: Xwin-LM/Xwin-LM-70B-V0.1
- sources:
  - layer_range:
    - 8
    - 9
    model: ehartford/Samantha-1.11-70b
- sources:
  - layer_range:
    - 9
    - 10
    model: Sao10K/Euryale-1.3-L2-70B
- sources:
  - layer_range:
    - 10
    - 11
    model: ehartford/Samantha-1.11-70b
- sources:
  - layer_range:
    - 11
    - 12
    model: jondurbin/airoboros-l2-c70b-3.1.2
- sources:
  - layer_range:
    - 12
    - 13
    model: fangloveskari/ORCA_LLaMA_70B_QLoRA
- sources:
  - layer_range:
    - 13
    - 14
    model: Sao10K/Euryale-1.3-L2-70B
- sources:
  - layer_range:
    - 14
    - 15
    model: fangloveskari/ORCA_LLaMA_70B_QLoRA
- sources:
  - layer_range:
    - 15
    - 16
    model: Sao10K/Euryale-1.3-L2-70B
- sources:
  - layer_range:
    - 16
    - 17
    model: fangloveskari/ORCA_LLaMA_70B_QLoRA
- sources:
  - layer_range:
    - 17
    - 18
    model: jondurbin/airoboros-l2-c70b-3.1.2
- sources:
  - layer_range:
    - 18
    - 19
    model: NousResearch/Nous-Hermes-Llama2-70b
- sources:
  - layer_range:
    - 19
    - 20
    model: Xwin-LM/Xwin-LM-70B-V0.1
- sources:
  - layer_range:
    - 20
    - 21
    model: Sao10K/Euryale-1.3-L2-70B
- sources:
  - layer_range:
    - 21
    - 22
    model: ehartford/Samantha-1.11-70b
- sources:
  - layer_range:
    - 22
    - 23
    model: jondurbin/airoboros-l2-c70b-3.1.2
- sources:
  - layer_range:
    - 23
    - 24
    model: augtoma/qCammel-70-x
- sources:
  - layer_range:
    - 24
    - 25
    model: Sao10K/Euryale-1.3-L2-70B
- sources:
  - layer_range:
    - 25
    - 27
    model: jondurbin/airoboros-l2-c70b-3.1.2
- sources:
  - layer_range:
    - 27
    - 28
    model: Xwin-LM/Xwin-LM-70B-V0.1
- sources:
  - layer_range:
    - 28
    - 29
    model: ehartford/Samantha-1.11-70b
- sources:
  - layer_range:
    - 29
    - 30
    model: Sao10K/Euryale-1.3-L2-70B
- sources:
  - layer_range:
    - 30
    - 32
    model: Xwin-LM/Xwin-LM-70B-V0.1
- sources:
  - layer_range:
    - 32
    - 33
    model: ehartford/Samantha-1.11-70b
- sources:
  - layer_range:
    - 33
    - 34
    model: augtoma/qCammel-70-x
- sources:
  - layer_range:
    - 34
    - 35
    model: Xwin-LM/Xwin-LM-70B-V0.1
- sources:
  - layer_range:
    - 35
    - 37
    model: Sao10K/Euryale-1.3-L2-70B
- sources:
  - layer_range:
    - 37
    - 38
    model: jondurbin/airoboros-l2-c70b-3.1.2
- sources:
  - layer_range:
    - 38
    - 39
    model: ehartford/Samantha-1.11-70b
- sources:
  - layer_range:
    - 39
    - 40
    model: augtoma/qCammel-70-x
- sources:
  - layer_range:
    - 40
    - 41
    model: Xwin-LM/Xwin-LM-70B-V0.1
- sources:
  - layer_range:
    - 41
    - 42
    model: ehartford/Samantha-1.11-70b
- sources:
  - layer_range:
    - 42
    - 43
    model: Sao10K/Euryale-1.3-L2-70B
- sources:
  - layer_range:
    - 43
    - 44
    model: Xwin-LM/Xwin-LM-70B-V0.1
- sources:
  - layer_range:
    - 44
    - 45
    model: NousResearch/Nous-Hermes-Llama2-70b
- sources:
  - layer_range:
    - 45
    - 46
    model: jondurbin/airoboros-l2-c70b-3.1.2
- sources:
  - layer_range:
    - 46
    - 48
    model: ehartford/Samantha-1.11-70b
- sources:
  - layer_range:
    - 48
    - 49
    model: Sao10K/Euryale-1.3-L2-70B
- sources:
  - layer_range:
    - 49
    - 50
    model: Xwin-LM/Xwin-LM-70B-V0.1
- sources:
  - layer_range:
    - 50
    - 51
    model: jondurbin/airoboros-l2-c70b-3.1.2
- sources:
  - layer_range:
    - 51
    - 54
    model: fangloveskari/ORCA_LLaMA_70B_QLoRA
- sources:
  - layer_range:
    - 54
    - 55
    model: jondurbin/airoboros-l2-c70b-3.1.2
- sources:
  - layer_range:
    - 55
    - 56
    model: fangloveskari/ORCA_LLaMA_70B_QLoRA
- sources:
  - layer_range:
    - 56
    - 58
    model: jondurbin/airoboros-l2-c70b-3.1.2
- sources:
  - layer_range:
    - 58
    - 59
    model: Sao10K/Euryale-1.3-L2-70B
- sources:
  - layer_range:
    - 59
    - 60
    model: Xwin-LM/Xwin-LM-70B-V0.1
- sources:
  - layer_range:
    - 60
    - 62
    model: jondurbin/airoboros-l2-c70b-3.1.2
- sources:
  - layer_range:
    - 62
    - 63
    model: Xwin-LM/Xwin-LM-70B-V0.1
- sources:
  - layer_range:
    - 63
    - 64
    model: fangloveskari/ORCA_LLaMA_70B_QLoRA
- sources:
  - layer_range:
    - 64
    - 65
    model: NousResearch/Nous-Hermes-Llama2-70b
- sources:
  - layer_range:
    - 65
    - 66
    model: Sao10K/Euryale-1.3-L2-70B
- sources:
  - layer_range:
    - 66
    - 67
    model: Xwin-LM/Xwin-LM-70B-V0.1
- sources:
  - layer_range:
    - 67
    - 68
    model: augtoma/qCammel-70-x
- sources:
  - layer_range:
    - 68
    - 70
    model: Xwin-LM/Xwin-LM-70B-V0.1
- sources:
  - layer_range:
    - 70
    - 71
    model: augtoma/qCammel-70-x
- sources:
  - layer_range:
    - 71
    - 72
    model: Xwin-LM/Xwin-LM-70B-V0.1
- sources:
  - layer_range:
    - 72
    - 73
    model: Sao10K/Euryale-1.3-L2-70B
- sources:
  - layer_range:
    - 73
    - 75
    model: jondurbin/airoboros-l2-c70b-3.1.2
- sources:
  - layer_range:
    - 75
    - 76
    model: Sao10K/Euryale-1.3-L2-70B
- sources:
  - layer_range:
    - 76
    - 77
    model: augtoma/qCammel-70-x
- sources:
  - layer_range:
    - 77
    - 78
    model: Xwin-LM/Xwin-LM-70B-V0.1
- sources:
  - layer_range:
    - 78
    - 79
    model: NousResearch/Nous-Hermes-Llama2-70b
- sources:
  - layer_range:
    - 79
    - 80
    model: Xwin-LM/Xwin-LM-70B-V0.1


=========================

=> Applying Doctor-Shotgun/limarpv3-llama2-70b-qlora x 0.35

Prompt template: Alpaca

Below is an instruction that describes a task. Write a response that appropriately completes the request.

### Instruction:
{prompt}

### Response:

A big thanks to Charles for adding the layer shuffle method to his tool mergekit and Henky/KoboldAI for the machine he let me use.

If you want to support me, you can here.

Downloads last month
690
Safetensors
Model size
69B params
Tensor type
FP16
·
Inference Examples
This model does not have enough activity to be deployed to Inference API (serverless) yet. Increase its social visibility and check back later, or deploy to Inference Endpoints (dedicated) instead.

Model tree for Undi95/Dawn-v2-70B

Quantizations
5 models