luigi86 commited on
Commit
c56018a
1 Parent(s): 551a374

Upload README.md with huggingface_hub

Browse files
Files changed (1) hide show
  1. README.md +55 -0
README.md ADDED
@@ -0,0 +1,55 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ license: other
3
+ license_name: tongyi-qianwen
4
+ license_link: https://huggingface.co/Qwen/Qwen2-72B-Instruct/blob/main/LICENSE
5
+ language:
6
+ - en
7
+ - zh
8
+ pipeline_tag: text-generation
9
+ tags:
10
+ - chat
11
+ ---
12
+
13
+ # Roleplay Quantization in EXL2 format for Magnum v1
14
+
15
+ Quantized using the cleaned PIPPA roleplay dataset. Uploading as I didn't see anyone else do this one yet.
16
+
17
+ # Original Model card
18
+
19
+ ![](https://files.catbox.moe/ngqnb1.png)
20
+
21
+ This is the first in a series of models designed to replicate the prose quality of the Claude 3 models, specifically Sonnet and Opus. This model is fine-tuned on top of [Qwen-2 72B Instruct](https://huggingface.co/Qwen/Qwen2-72B-Instruct).
22
+
23
+
24
+ ## Prompting
25
+ Model has been Instruct tuned with the ChatML formatting. A typical input would look like this:
26
+
27
+ ```py
28
+ """<|im_start|>user
29
+ Hi there!<|im_end|>
30
+ <|im_start|>assistant
31
+ Nice to meet you!<|im_end|>
32
+ <|im_start|>user
33
+ Can I ask a question?<|im_end|>
34
+ <|im_start|>assistant
35
+ """
36
+ ```
37
+
38
+ ## Credits
39
+
40
+ This model has been a team effort, credits go to:
41
+
42
+ - [Sao10K](https://huggingface.co/Sao10K) for help with (and cleaning up!) the dataset.
43
+ - [alpindale](https://huggingface.co/alpindale) for the training.
44
+ - [kalomaze](https://huggingface.co/kalomaze) for helping with the hyperparameter tuning.
45
+ - Various other people for their continued help as we tuned the parameters, restarted failed runs. In no particular order: [Doctor Shotgun](https://huggingface.co/Doctor-Shotgun), [Lucy](https://huggingface.co/lucyknada), [Nopm](https://huggingface.co/nopm), [Mango](https://huggingface.co/MangoMango69420), and the rest of the Silly Tilly.
46
+
47
+ And last but not least, we'd like to thank [Kearm](https://twitter.com/Nottlespike) for sponsoring the compute needed to train this model.
48
+
49
+ ## Training
50
+ The training was done with 55 million tokens of high-quality RP data, over 1.5 epochs. We used 8x [AMD Instinct™ MI300X Accelerators](https://www.amd.com/en/products/accelerators/instinct/mi300/mi300x.html) for the full-parameter fine-tuning of the model.
51
+
52
+ [<img src="https://raw.githubusercontent.com/OpenAccess-AI-Collective/axolotl/main/image/axolotl-badge-web.png" alt="Built with Axolotl" width="200" height="32"/>](https://github.com/OpenAccess-AI-Collective/axolotl)
53
+
54
+ ## Safety
55
+ ...