Triangle104 commited on
Commit
5214608
·
verified ·
1 Parent(s): 77322c5

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +53 -0
README.md CHANGED
@@ -15,6 +15,59 @@ base_model: anthracite-org/magnum-v4-9b
15
  This model was converted to GGUF format from [`anthracite-org/magnum-v4-9b`](https://huggingface.co/anthracite-org/magnum-v4-9b) using llama.cpp via the ggml.ai's [GGUF-my-repo](https://huggingface.co/spaces/ggml-org/gguf-my-repo) space.
16
  Refer to the [original model card](https://huggingface.co/anthracite-org/magnum-v4-9b) for more details on the model.
17
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
18
  ## Use with llama.cpp
19
  Install llama.cpp through brew (works on Mac and Linux)
20
 
 
15
  This model was converted to GGUF format from [`anthracite-org/magnum-v4-9b`](https://huggingface.co/anthracite-org/magnum-v4-9b) using llama.cpp via the ggml.ai's [GGUF-my-repo](https://huggingface.co/spaces/ggml-org/gguf-my-repo) space.
16
  Refer to the [original model card](https://huggingface.co/anthracite-org/magnum-v4-9b) for more details on the model.
17
 
18
+ ---
19
+ This is a series of models designed to replicate the prose quality of the Claude 3 models, specifically Sonnet and Opus.
20
+
21
+ This model is fine-tuned on top of gemma 2 9b (chatML'ified).
22
+ Prompting
23
+
24
+ A typical input would look like this:
25
+
26
+ <|im_start|>system
27
+ system prompt<|im_end|>
28
+ <|im_start|>user
29
+ Hi there!<|im_end|>
30
+ <|im_start|>assistant
31
+ Nice to meet you!<|im_end|>
32
+ <|im_start|>user
33
+ Can I ask a question?<|im_end|>
34
+ <|im_start|>assistant
35
+
36
+ SillyTavern templates
37
+
38
+ Below are Instruct and Context templates for use within SillyTavern.
39
+ context template
40
+
41
+
42
+ instruct template
43
+
44
+
45
+ Axolotl config
46
+ See axolotl config
47
+
48
+
49
+ Credits
50
+
51
+ We'd like to thank Recursal / Featherless for sponsoring the compute for this train, Featherless has been hosting our Magnum models since the first 72 B and has given thousands of people access to our models and helped us grow.
52
+
53
+ We would also like to thank all members of Anthracite who made this finetune possible.
54
+ Datasets
55
+
56
+ anthracite-org/c2_logs_16k_llama_v1.1
57
+ NewEden/Claude-Instruct-5K
58
+ anthracite-org/kalo-opus-instruct-22k-no-refusal
59
+ Epiculous/SynthRP-Gens-v1.1-Filtered-n-Cleaned
60
+ lodrick-the-lafted/kalo-opus-instruct-3k-filtered
61
+ anthracite-org/nopm_claude_writing_fixed
62
+ Epiculous/Synthstruct-Gens-v1.1-Filtered-n-Cleaned
63
+ anthracite-org/kalo_opus_misc_240827
64
+ anthracite-org/kalo_misc_part2
65
+
66
+ Training
67
+
68
+ The training was done for 2 epochs. We used 8xH100s GPUs graciously provided by Recursal AI / Featherless AI for the full-parameter fine-tuning of the model.
69
+
70
+ ---
71
  ## Use with llama.cpp
72
  Install llama.cpp through brew (works on Mac and Linux)
73