--- license: other --- See LICENSE file for license. This is a collection of merged, then converted to 4bit LLaMA models trained on the storytelling dataset I used for the storytelling LoRAs. UPDATE: 04/04 Cleaned data and retrained to 32 groupsize and safetensors. Formatting oddities seem to have been wiped out. Format: Nothing notable, chapters separated by *** therefore may mess some things up. UPDATE: 2024-04-18 Retrained and merged using updated LoRAs. To merge and convert, used: ``` transformers 4.28.1. gptq cuda branch 5731aa1 llamacpp master branch 8944a13 ``` Notes for usage. ``` - These models are not instruct LoRAs. They are designed to supplement existing story data. - There will likely be some bleedthrough on locations and names, this is especially notable if you use with very little context. - There isn't any large notable formatting, ### seperated stories in the dataset, and *** seperated chapters. ``` Currently transferring models over. ``` 7B safetensors 4bit - 7B ggml 4bit - 13B safetensors 4bit - 13B ggml 4bit - 30B safetensors 4bit - 30B ggml 4bit - ```