Update README.md
Browse files
README.md
CHANGED
@@ -29,10 +29,16 @@ Can I ask a question?<|im_end|>
|
|
29 |
|
30 |
## Credits
|
31 |
|
|
|
|
|
|
|
|
|
|
|
|
|
32 |
This model has been a team effort, and the credits goes to all members of Anthracite.
|
33 |
|
34 |
## Training
|
35 |
-
The training was done with a
|
36 |
|
37 |
[<img src="https://raw.githubusercontent.com/OpenAccess-AI-Collective/axolotl/main/image/axolotl-badge-web.png" alt="Built with Axolotl" width="200" height="32"/>](https://github.com/OpenAccess-AI-Collective/axolotl)
|
38 |
|
|
|
29 |
|
30 |
## Credits
|
31 |
|
32 |
+
Three new general purpose instruction following datasets were added on top of the original Stheno dataset (which had certain low quality entries purged/removed).
|
33 |
+
The first two were designed specifically for the Magnum series, to better address prompt adherence and coherence:
|
34 |
+
- [kalomaze/Opus_Instruct_25k](https://huggingface.co/datasets/kalomaze/Opus_Instruct_25k)
|
35 |
+
- [Nopm/Opus_WritingStruct](https://huggingface.co/datasets/Nopm/Opus_WritingStruct)
|
36 |
+
- [Gryphe/Sonnet3.5-SlimOrcaDedupCleaned](https://huggingface.co/datasets/Gryphe/Sonnet3.5-SlimOrcaDedupCleaned) (A ~16k rows subset)
|
37 |
+
|
38 |
This model has been a team effort, and the credits goes to all members of Anthracite.
|
39 |
|
40 |
## Training
|
41 |
+
The training was done for 2 epochs with a learning rate of 1e-05. We used 8x [NVIDIA H100 Tensor Core](https://www.nvidia.com/en-us/data-center/h100/) GPUs for the full-parameter fine-tuning of the model.
|
42 |
|
43 |
[<img src="https://raw.githubusercontent.com/OpenAccess-AI-Collective/axolotl/main/image/axolotl-badge-web.png" alt="Built with Axolotl" width="200" height="32"/>](https://github.com/OpenAccess-AI-Collective/axolotl)
|
44 |
|