File size: 1,449 Bytes
b1cd6bb f587bd1 b1cd6bb 8d04773 1aafa0f 8d04773 7ddc1e0 8d04773 e43bf2d |
1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30 |
---
license: cc-by-sa-4.0
datasets:
- acrastt/EverythingLM-V3-ShareGPT
language:
- en
library_name: transformers
pipeline_tag: text-generation
---
<a href="https://www.buymeacoffee.com/acrastt" target="_blank"><img src="https://cdn.buymeacoffee.com/buttons/v2/default-yellow.png" alt="Buy Me A Coffee" style="height: 60px !important;width: 217px !important;" ></a>
This is [StableLM 3B 4E1T](https://huggingface.co/stabilityai/stablelm-3b-4e1t) finetuned on [EverythingLM Data V3(ShareGPT Format)](https://huggingface.co/datasets/acrastt/EverythingLM-V3-ShareGPT) for 2 epochs using [QLoRA](https://arxiv.org/abs/2305.14314).
Prompt template:
```
### HUMAN:
{prompt}
### RESPONSE:
```
Note that this model have the EOS token of `<|endoftext|>` instead of `<\s>`.
## Attribution
[StableLM 3B 4E1T](https://huggingface.co/stabilityai/stablelm-3b-4e1t) by [Stability AI](https://stability.ai/) is licensed under [CC BY-SA 4.0](https://creativecommons.org/licenses/by-sa/4.0/).
Modifications:
- Instruction tuned on dataset [EverythingLM Data V3(ShareGPT Format)](https://huggingface.co/datasets/acrastt/EverythingLM-V3-ShareGPT) for 2 epochs using [QLoRA](https://arxiv.org/abs/2305.14314).
- Added `"_name_or_path"`, changed `"torch_dtype"` and `"transformers_version"` in [`config.json`](https://huggingface.co/acrastt/Marx-3B-V3/blob/main/config.json).
- Changes to [`README.MD`](https://huggingface.co/acrastt/Marx-3B-V3/blob/main/README.md). |