File size: 1,814 Bytes
0d37aca |
1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 |
[<img src="https://raw.githubusercontent.com/OpenAccess-AI-Collective/axolotl/main/image/axolotl-badge-web.png" alt="Built with Axolotl" width="200" height="32"/>](https://github.com/OpenAccess-AI-Collective/axolotl)
**[💵 Donate to OpenAccess AI Collective](https://github.com/sponsors/OpenAccess-AI-Collective) to help us keep building great tools and models!**
# Dodona Pyg 15B 8K Preview
This finetune adds about 100MB of the v8p4 dataset to Dodona 15B 8K.
Dodona 15B 8K Preview is an experiment for fan-fiction and character ai use cases. It is built on Starcoder Plus to give it 8K context length and pretrained on a corpus of fanfiction and visual novels.
Lots of mistakes were made during the creation of this model, but we didn't want to throw $300 of model training time out the window, so we are releasing this as a preview.
If you would like to see us continue to build more models like this, please consider donating by sponsoring us on GitHub on the link above or [Buy me a coffee](https://www.buymeacoffee.com/winglian).
Questions, comments, feedback, looking to donate, or want to help? Reach out on our [Discord](https://discord.gg/PugNNHAF5r) or email [wing@openaccessaicollective.org](mailto:wing@openaccessaicollective.org)
# Prompts
While this model is minimally finetuned with USER: / ASSISTANT: prompts, it seems to respond better to Alpaca style prompts:
```
Below is an instruction that describes a task. Write a response that appropriately completes the request.
### Instruction:
...
### Response:
```
<img src="https://huggingface.co/openaccess-ai-collective/dodona-15b-preview/resolve/main/dodona.png" alt="oracle of dodona" width="600" height="500"/>
|