Safetensors
gemma2
8-bit precision
exl2
File size: 1,988 Bytes
594a9fe
 
 
 
 
 
 
 
 
 
 
 
7a3627d
594a9fe
609f71c
 
 
594a9fe
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
---
license: agpl-3.0
datasets:
- kalomaze/Opus_Instruct_25k
- kalomaze/Opus_Instruct_3k
- MangoHQ/Claude-Data-Anon-Killed
---



![image/png](https://cdn-uploads.huggingface.co/production/uploads/66209411a245d5aa78585f88/hYuzluOiES5cedXGEQ0_J.png)

GGUF: [Here](https://huggingface.co/MangyMango/Ohashi-9B-V1-GGUF)

EXL2 6.0BPW (Thx Lucy <3): [Here](https://huggingface.co/MangyMango/Ohashi-9B-V1-EXl2-6.0BPW)

More quants will be up soon.


# Model Details

I took Gemma-2 base and trained a LoRa with 2 million tokens worth of Claude data and merged via Axolotl CLI 
Data used was similar to what Magnum Models are trained off hence Claude Shannon for the card image.

# Prompting

In testing it worked well with basic sampler settings (specifically i used Simple~1 included within ST); it was coherent and stable throughout my testing aswell as being quite proactive. I used the Gemma2 format provided in SillyTavern to test and i found no refusals within RP even when doing extreme NSFW activites - When i was using it as an assistant though i found many refusals but all of them were easily dealt with by using MooreRP, a custom prompt / context template to uncensor the model

MooreRP links: 
Context Template: https://files.catbox.moe/b1lpao.json
Instruct Mode: https://files.catbox.moe/21joxa.json

(Made by @a.lk on Discord)

# Config

LoRa for this model was trained in Axoltol for 2 epochs at a rank of 32 and a LR of 2e-5 on 2xRTX 6000s (Provided by Kubernetes Bad) and using the [CustomGemma2 Prompt strategy](https://github.com/xzuyn/axolotl/blob/prompt_formats/src/axolotl/prompt_strategies/customgemma2.py)

# Credits

Thanks to [Kubernetes Bad](https://huggingface.co/kubernetes-bad) for providing compute for this train, [Lucy Knada](https://huggingface.co/lucyknada), [Nopm](https://huggingface.co/Nopm), [Kalomaze](https://huggingface.co/kalomaze) and the rest of [Anthracite](https://huggingface.co/anthracite-org) for providing help to do the train. (But not Alpin)