Text Generation
Safetensors
mistral
Merge
conversational
Epiculous commited on
Commit
c2a16a9
1 Parent(s): db8d52b

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +36 -24
README.md CHANGED
@@ -1,32 +1,48 @@
1
  ---
2
- base_model:
3
- - Epiculous/Crimson_Dawn-V0.1
4
- - Epiculous/Azure_Dusk-v0.1
5
- library_name: transformers
 
 
 
 
 
 
 
 
 
 
 
6
  tags:
7
- - mergekit
8
  - merge
9
-
10
  ---
11
- # Violet_Twilight
12
-
13
- This is a merge of pre-trained language models created using [mergekit](https://github.com/cg123/mergekit).
14
-
15
- ## Merge Details
16
- ### Merge Method
17
 
18
- This model was merged using the SLERP merge method.
19
 
20
- ### Models Merged
21
 
22
- The following models were included in the merge:
23
- * [Epiculous/Crimson_Dawn-V0.1](https://huggingface.co/Epiculous/Crimson_Dawn-V0.1)
24
- * [Epiculous/Azure_Dusk-v0.1](https://huggingface.co/Epiculous/Azure_Dusk-v0.1)
25
 
26
- ### Configuration
 
27
 
28
- The following YAML configuration was used to produce this model:
 
 
 
 
 
 
 
 
 
 
 
29
 
 
 
30
  ```yaml
31
  slices:
32
  - sources:
@@ -34,10 +50,6 @@ slices:
34
  layer_range: [0, 40]
35
  - model: Epiculous/Crimson_Dawn-V0.1
36
  layer_range: [0, 40]
37
- # or, the equivalent models: syntax:
38
- # models:
39
- # - model: psmathur/orca_mini_v3_13b
40
- # - model: garage-bAInd/Platypus2-13B
41
  merge_method: slerp
42
  base_model: Epiculous/Azure_Dusk-v0.1
43
  parameters:
@@ -49,4 +61,4 @@ parameters:
49
  - value: 0.5 # fallback for rest of tensors
50
  dtype: bfloat16
51
 
52
- ```
 
1
  ---
2
+ license: apache-2.0
3
+ datasets:
4
+ - Epiculous/SynthRP-Gens-v1-Filtered-n-Cleaned
5
+ - Epiculous/Synthstruct-Gens-v1-Filtered-n-Cleaned
6
+ language:
7
+ - en
8
+ - fr
9
+ - de
10
+ - es
11
+ - it
12
+ - pt
13
+ - ru
14
+ - zh
15
+ - ja
16
+ pipeline_tag: text-generation
17
  tags:
 
18
  - merge
 
19
  ---
 
 
 
 
 
 
20
 
21
+ ![image/png](https://cdn-uploads.huggingface.co/production/uploads/64adfd277b5ff762771e4571/oTCM28M_Hu6lM4DmOp7tC.png)
22
 
23
+ Now for something a bit different, Violet_Twilight! This model is a SLERP merge of Azure_Dusk and Crimson_Dawn!
24
 
25
+ # Quants!
26
+ <strong>full</strong> / [exl2](https://huggingface.co/Epiculous/Violet_Twilight-v0.1-Exl2) / [gguf](https://huggingface.co/Epiculous/Violet_Twilight-v0.1-GGUF)
 
27
 
28
+ ## Prompting
29
+ Violet_Twilight's models were trained with the Mistral Instruct template, therefore it should be prompted in a similar way that you would prompt any other mistral based model.
30
 
31
+ ```
32
+ "<s>[INST] Prompt goes here [/INST]<\s>"
33
+ ```
34
+ ### Context and Instruct
35
+ [Magnum-123B-Context.json](https://files.catbox.moe/rkyqwg.json) <br/>
36
+ [Magnum-123B-Instruct.json](https://files.catbox.moe/obb5oe.json) <br/>
37
+ *** NOTE *** <br/>
38
+ There have been reports of the quantized model misbehaving with the mistral prompt, if you are seeing issues it may be worth trying ChatML Context and Instruct templates.
39
+ If you are using GGUF I strongly advise using ChatML, for some reason that quantization performs better using ChatML.
40
+ ### Current Top Sampler Settings
41
+ [Crimson_Dawn-Nitral-Special](https://files.catbox.moe/8xjxht.json) - Considered the best settings! <br/>
42
+ [Crimson_Dawn-Magnum-Style](https://files.catbox.moe/lc59dn.json)
43
 
44
+ ## Merging
45
+ The following config was used to merge Azure Dusk and Crimson Dawn
46
  ```yaml
47
  slices:
48
  - sources:
 
50
  layer_range: [0, 40]
51
  - model: Epiculous/Crimson_Dawn-V0.1
52
  layer_range: [0, 40]
 
 
 
 
53
  merge_method: slerp
54
  base_model: Epiculous/Azure_Dusk-v0.1
55
  parameters:
 
61
  - value: 0.5 # fallback for rest of tensors
62
  dtype: bfloat16
63
 
64
+ ```