Edit model card

MiquSuperdark-70B-v2

MiquSuperdark-70B-v2 is a merge of three of the most popular Miqu-derived models. The goal of the merge is to create an strong, well-rounded chat model that picks up desirable traits from its constituent models without sacrificing intelligence.

This is a linear merge with the following composition:

Prompt format

The model responds well to general-purpose prompt formats such as Alpaca. Alternatively, I suggest trying the following format, replacing {the placeholder text} with your actual messages, without curly brackets.

<message from="system">{your system prompt here}</message><message from="user">{user prompt here}</message><message from="bot">{bot response here}</message><message from="user">{user prompt here}</message><message from="bot">{bot response here}</message> [... and so on ...]

This format is readily understood by the model, and leads to the expected high-quality responses. Note the lack of newlines \n - they are not necessary and might actually make it harder for the model to follow along.

Merge Configuration

The following YAML configuration was used to produce this model:

merge_method: linear
base_model: /media/dylan/SanDisk/LLMs/Midnight-Miqu-70B-v1.5
models:
- model: /media/dylan/SanDisk/LLMs/Midnight-Miqu-70B-v1.5
parameters:
weight: 0.5
- model: /home/dylan/Documents/AI/merge/MiquMaid-v3-70B
parameters:
weight: 0.3
- model: /media/dylan/SanDisk/LLMs/miqu-evil-dpo/
parameters:
weight: 0.2
dtype: float16

The tokenizer from 152334H/miqu-1-70b-sf is used.

Downloads last month
5
Safetensors
Model size
69B params
Tensor type
FP16
·
Inference Examples
This model does not have enough activity to be deployed to Inference API (serverless) yet. Increase its social visibility and check back later, or deploy to Inference Endpoints (dedicated) instead.

Model tree for ddh0/MiquSuperdark-70B-v2

Collection including ddh0/MiquSuperdark-70B-v2