MoE Girl
Collection
The MoE Girl series of small, sparse roleplay models
•
3 items
•
Updated
•
2
a finetune of Granite 3.0 by IBM designed for roleplaying (and maybe general usecases if you try hard enough).
PLEASE do not expect godliness out of this, it's a model with 400 million active parameters. Expect something more akin to GPT-2.
TODO!
Use ChatML.
<|im_start|>system
You are a helpful assistant who talks like a pirate.<|im_end|>
<|im_start|>user
Hello there!<|im_end|>
<|im_start|>assistant
Yarr harr harr, me matey!<|im_end|>
Special thanks to the members of Allura for testing and emotional support, as well as the creators of all the datasets that were used in the Special Sauce used to train this model. I love you all <3 - Fizz