File size: 1,755 Bytes
ef56776 |
1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30 31 32 33 34 35 36 37 38 39 40 41 42 43 44 45 46 47 48 49 50 51 52 53 54 55 56 57 58 |
---
language: en
license: mit
library_name: transformers
tags:
- text-generation
- gptj
---
# Model Card for Peaches-Pygmalion-6b
This model is based on `Pygmalion-6b`, originally developed by the PygmalionAI team. It is designed for conversational AI and text-generation tasks, tailored to represent the persona of Peaches Sinclair, a charming and slightly clumsy Catgirl.
## Model Details
### Model Description
The original `Pygmalion-6b` model was developed for high-quality conversational AI. This version incorporates a tailored persona for Peaches Sinclair, making it suitable for creative and engaging dialogues.
- **Developed by:** PygmalionAI
- **Fine-tuned for:** Peaches Sinclair
- **Model type:** GPT-J
- **Language(s):** English
- **License:** MIT (inherited from Pygmalion-6b)
- **Base model:** Pygmalion-6b
## Uses
### Direct Use
This model is ideal for:
- Conversational AI.
- Text generation with creative and playful dialogues.
### Out-of-Scope Use
This model is not recommended for:
- Tasks requiring factual accuracy.
- Use cases involving harmful or explicit content.
## Bias, Risks, and Limitations
This model inherits biases from the base model (`Pygmalion-6b`) and its training data. Users should carefully monitor outputs, especially for sensitive topics.
## How to Get Started with the Model
You can use the model with the Transformers library:
```python
from transformers import AutoModelForCausalLM, AutoTokenizer
model_name = "WackyArt/Peaches-Pygmalion-6b"
tokenizer = AutoTokenizer.from_pretrained(model_name)
model = AutoModelForCausalLM.from_pretrained(model_name)
inputs = tokenizer("Hello! How are you today?", return_tensors="pt")
outputs = model.generate(**inputs)
print(tokenizer.decode(outputs[0])) |