Edit model card

Borealis

image/png

Borealis-10.7B-DPO is a 10.7B model made of 48 Mistral 7B layers, finetuned for +70h on 2xA6000 on a big RP and Conversational dataset with llama2 configuration of Axolotl, like SOLAR.

This variant had a DPO train on top of it.

Description

This repo contains fp16 files of Borealis-10.7B-DPO, a conversational model.

The goal of this model isn't to break all benchmark, but to have a better RP/ERP/Conversational model.

It was trained on multiple basic dataset to make it intelligent, but majority of the dataset was basic conversations.

Dataset used

  • NobodyExistsOnTheInternet/ToxicQAFinal
  • teknium/openhermes
  • unalignment/spicy-3.1
  • Doctor-Shotgun/no-robots-sharegpt
  • Undi95/toxic-dpo-v0.1-sharegpt
  • Aesir [1], [2], [3-SFW], [3-NSFW]
  • lemonilia/LimaRP
  • Squish42/bluemoon-fandom-1-1-rp-cleaned
  • Undi95/ConversationChronicles-sharegpt-SHARDED (2 sets, modified)

DPO Dataset used

  • Intel/orca_dpo_pairs
  • NobodyExistsOnTheInternet/ToxicDPOqa
  • Undi95/toxic-dpo-v0.1-NoWarning

Prompt format: NsChatml

<|im_system|>
{sysprompt}<|im_end|>
<|im_user|>
{input}<|im_end|>
<|im_bot|>
{output}<|im_end|>

Others

If you want to support me, you can here.

Downloads last month
645
Inference Examples
This model does not have enough activity to be deployed to Inference API (serverless) yet. Increase its social visibility and check back later, or deploy to Inference Endpoints (dedicated) instead.

Model tree for Undi95/Borealis-10.7B-DPO

Merges
1 model
Quantizations
2 models