Fimbul-Airo-18B / README.md
Burn Oil
Update README.md
50531dc verified
|
raw
history blame
1.56 kB
metadata
base_model:
  - Sao10K/Fimbulvetr-11B-v2
  - Undi95/Mistral-11B-CC-Air-RP
library_name: transformers
tags:
  - mergekit
  - merge
  - πŸ‘

Fimbul-Airo-18B πŸ‘

This is a merge of pre-trained language models created using mergekit. πŸ‘

I tested it for thirtneen.second πŸ‘

Works pretty good, seems uncensored. I'll update with more results/observations as I continue to test.

Merge Details

Merge Method

This model was merged using the passthrough merge method. Taking models and smashing em all together πŸ‘

Models Merged

The following models were included in the merge:

The Sauce

The following YAML configuration was used to produce this model:

slices:
  - sources:
    - model: Sao10K/Fimbulvetr-11B-v2
      layer_range: [0, 40]
  - sources:
    - model: Undi95/Mistral-11B-CC-Air-RP
      layer_range: [8, 48]
merge_method: passthrough
dtype: bfloat16

πŸ‘

Prompt Format: Alpaca πŸ‘

### Instruction:
<Prompt>

### Input:
<Insert Context Here>

### Response:

πŸ‘

Don't forget to take care of yourself and have a wonderful day!