Update README.md
Browse files
README.md
CHANGED
@@ -15,6 +15,59 @@ base_model: anthracite-org/magnum-v4-9b
|
|
15 |
This model was converted to GGUF format from [`anthracite-org/magnum-v4-9b`](https://huggingface.co/anthracite-org/magnum-v4-9b) using llama.cpp via the ggml.ai's [GGUF-my-repo](https://huggingface.co/spaces/ggml-org/gguf-my-repo) space.
|
16 |
Refer to the [original model card](https://huggingface.co/anthracite-org/magnum-v4-9b) for more details on the model.
|
17 |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
18 |
## Use with llama.cpp
|
19 |
Install llama.cpp through brew (works on Mac and Linux)
|
20 |
|
|
|
15 |
This model was converted to GGUF format from [`anthracite-org/magnum-v4-9b`](https://huggingface.co/anthracite-org/magnum-v4-9b) using llama.cpp via the ggml.ai's [GGUF-my-repo](https://huggingface.co/spaces/ggml-org/gguf-my-repo) space.
|
16 |
Refer to the [original model card](https://huggingface.co/anthracite-org/magnum-v4-9b) for more details on the model.
|
17 |
|
18 |
+
---
|
19 |
+
This is a series of models designed to replicate the prose quality of the Claude 3 models, specifically Sonnet and Opus.
|
20 |
+
|
21 |
+
This model is fine-tuned on top of gemma 2 9b (chatML'ified).
|
22 |
+
Prompting
|
23 |
+
|
24 |
+
A typical input would look like this:
|
25 |
+
|
26 |
+
<|im_start|>system
|
27 |
+
system prompt<|im_end|>
|
28 |
+
<|im_start|>user
|
29 |
+
Hi there!<|im_end|>
|
30 |
+
<|im_start|>assistant
|
31 |
+
Nice to meet you!<|im_end|>
|
32 |
+
<|im_start|>user
|
33 |
+
Can I ask a question?<|im_end|>
|
34 |
+
<|im_start|>assistant
|
35 |
+
|
36 |
+
SillyTavern templates
|
37 |
+
|
38 |
+
Below are Instruct and Context templates for use within SillyTavern.
|
39 |
+
context template
|
40 |
+
|
41 |
+
|
42 |
+
instruct template
|
43 |
+
|
44 |
+
|
45 |
+
Axolotl config
|
46 |
+
See axolotl config
|
47 |
+
|
48 |
+
|
49 |
+
Credits
|
50 |
+
|
51 |
+
We'd like to thank Recursal / Featherless for sponsoring the compute for this train, Featherless has been hosting our Magnum models since the first 72 B and has given thousands of people access to our models and helped us grow.
|
52 |
+
|
53 |
+
We would also like to thank all members of Anthracite who made this finetune possible.
|
54 |
+
Datasets
|
55 |
+
|
56 |
+
anthracite-org/c2_logs_16k_llama_v1.1
|
57 |
+
NewEden/Claude-Instruct-5K
|
58 |
+
anthracite-org/kalo-opus-instruct-22k-no-refusal
|
59 |
+
Epiculous/SynthRP-Gens-v1.1-Filtered-n-Cleaned
|
60 |
+
lodrick-the-lafted/kalo-opus-instruct-3k-filtered
|
61 |
+
anthracite-org/nopm_claude_writing_fixed
|
62 |
+
Epiculous/Synthstruct-Gens-v1.1-Filtered-n-Cleaned
|
63 |
+
anthracite-org/kalo_opus_misc_240827
|
64 |
+
anthracite-org/kalo_misc_part2
|
65 |
+
|
66 |
+
Training
|
67 |
+
|
68 |
+
The training was done for 2 epochs. We used 8xH100s GPUs graciously provided by Recursal AI / Featherless AI for the full-parameter fine-tuning of the model.
|
69 |
+
|
70 |
+
---
|
71 |
## Use with llama.cpp
|
72 |
Install llama.cpp through brew (works on Mac and Linux)
|
73 |
|