Delta-Vector
commited on
Commit
•
80b8eea
1
Parent(s):
cc9d98e
Update README.md
Browse files
README.md
CHANGED
@@ -16,10 +16,9 @@ tags:
|
|
16 |
- chat
|
17 |
---
|
18 |
|
19 |
-
![image/png](https://huggingface.co/Edens-Gate/Testing123/resolve/main/oie_gM9EsNXjMDsT.jpg?download=true)
|
20 |
A model made to continue off my previous work on [Magnum 4B](https://huggingface.co/anthracite-org/magnum-v2-4b), A small model made for creative writing / General assistant tasks, finetuned ontop of [IntervitensInc/Llama-3.1-Minitron-4B-Width-Base-chatml](https://huggingface.co/IntervitensInc/Llama-3.1-Minitron-4B-Width-Base-chatml), this model is made to be more coherent and generally be better then the 4B at both writing and assistant tasks.
|
21 |
|
22 |
-
# EXL2 quants of Holland 4B,
|
23 |
|
24 |
|
25 |
## Prompting
|
@@ -160,7 +159,6 @@ special_tokens:
|
|
160 |
- [Epiculous/Synthstruct-Gens-v1.1-Filtered-n-Cleaned](https://huggingface.co/datasets/Epiculous/Synthstruct-Gens-v1.1-Filtered-n-Cleaned)
|
161 |
- [lodrick-the-lafted/OpusStories](https://huggingface.co/datasets/lodrick-the-lafted/OpusStories)
|
162 |
|
163 |
-
I couldn't have made this model without the help of [Kubernetes_bad](https://huggingface.co/kubernetes-bad) and the support of [Lucy Knada](https://huggingface.co/lucyknada)
|
164 |
|
165 |
## Training
|
166 |
The training was done for 2 epochs. We used 2 x [RTX 6000s](https://store.nvidia.com/en-us/nvidia-rtx/products/nvidia-rtx-6000-ada-generation/) GPUs graciously provided by [Kubernetes_Bad](https://huggingface.co/kubernetes-bad) for the full-parameter fine-tuning of the model.
|
|
|
16 |
- chat
|
17 |
---
|
18 |
|
|
|
19 |
A model made to continue off my previous work on [Magnum 4B](https://huggingface.co/anthracite-org/magnum-v2-4b), A small model made for creative writing / General assistant tasks, finetuned ontop of [IntervitensInc/Llama-3.1-Minitron-4B-Width-Base-chatml](https://huggingface.co/IntervitensInc/Llama-3.1-Minitron-4B-Width-Base-chatml), this model is made to be more coherent and generally be better then the 4B at both writing and assistant tasks.
|
20 |
|
21 |
+
# EXL2 quants of Holland 4B, Original weights can be found [here](https://huggingface.co/NewEden/Holland-4B)
|
22 |
|
23 |
|
24 |
## Prompting
|
|
|
159 |
- [Epiculous/Synthstruct-Gens-v1.1-Filtered-n-Cleaned](https://huggingface.co/datasets/Epiculous/Synthstruct-Gens-v1.1-Filtered-n-Cleaned)
|
160 |
- [lodrick-the-lafted/OpusStories](https://huggingface.co/datasets/lodrick-the-lafted/OpusStories)
|
161 |
|
|
|
162 |
|
163 |
## Training
|
164 |
The training was done for 2 epochs. We used 2 x [RTX 6000s](https://store.nvidia.com/en-us/nvidia-rtx/products/nvidia-rtx-6000-ada-generation/) GPUs graciously provided by [Kubernetes_Bad](https://huggingface.co/kubernetes-bad) for the full-parameter fine-tuning of the model.
|