Create README.md
Browse files
README.md
ADDED
@@ -0,0 +1,54 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
### Chatty-2x8B
|
2 |
+
|
3 |
+
[picture here]
|
4 |
+
|
5 |
+
## Description
|
6 |
+
After some testing, finetuning and multiple merges of Llama-3 LLM models, here is something a little different.
|
7 |
+
|
8 |
+
This model is a MoE of 2x Llama-3 model trained on different RP format.
|
9 |
+
|
10 |
+
This repo contains GGUF files of Chatty-2x8B.
|
11 |
+
|
12 |
+
## The idea
|
13 |
+
|
14 |
+
I started with two separate Llama-3-Instruct-8B models, each fine-tuned for specific RP formats.
|
15 |
+
|
16 |
+
Here is two simple exemple of how it was trained.
|
17 |
+
|
18 |
+
|
19 |
+
- **Expert 1**: This model is trained to handle RP that requires actions and descriptions between asterisks. For example:
|
20 |
+
```
|
21 |
+
*nods* Yes, I understand.
|
22 |
+
```
|
23 |
+
- **Expert 2**: This model is fine-tuned for plain text RP where characters’ dialogues and actions are described straightforwardly. For example:
|
24 |
+
```
|
25 |
+
Nods. "Yes, I understand."
|
26 |
+
```
|
27 |
+
|
28 |
+
My initial idea was to make a 11B or bigger Llama-3 model, or just make a 2x8B from existing model, but I got some issues, they were not stable enough, even after DPO and FFT on top my frankenmerge/moe of Llama-3, it was not working well enough to release them.
|
29 |
+
|
30 |
+
So I just tried the idea of having 2 different RP format trained on 2 separated Llama-3-Instruct-8B, and it worked pretty well!
|
31 |
+
|
32 |
+
## The dataset
|
33 |
+
|
34 |
+
Based on Lumimaid 8B OAS success I still used the same "balance" between RP and non RP in the dataset, the maximum was 50% non RP data on each side.
|
35 |
+
|
36 |
+
RP data was different with some exception, the non RP data was exactly the same, despite that, I can't produce repetition so the double usage of non RP datasets didn't hurt the model in the end.
|
37 |
+
|
38 |
+
## Prompt template: Llama3
|
39 |
+
|
40 |
+
```
|
41 |
+
<|begin_of_text|><|start_header_id|>system<|end_header_id|>
|
42 |
+
|
43 |
+
{system_prompt}<|eot_id|><|start_header_id|>user<|end_header_id|>
|
44 |
+
|
45 |
+
{input}<|eot_id|><|start_header_id|>assistant<|end_header_id|>
|
46 |
+
|
47 |
+
{output}<|eot_id|>
|
48 |
+
```
|
49 |
+
|
50 |
+
## Others
|
51 |
+
|
52 |
+
Undi: If you want to support us, you can [here](https://ko-fi.com/undiai).
|
53 |
+
|
54 |
+
IkariDev: Visit my [retro/neocities style website](https://ikaridevgit.github.io/) please kek
|