|
--- |
|
datasets: |
|
- unalignment/toxic-dpo-v0.1 |
|
--- |
|
|
|
# llama2_xs_460M_uncensored |
|
## Model Details |
|
|
|
[llama2_xs_460M_experimental](https://huggingface.co/ahxt/llama2_xs_460M_experimental) DPO finedtuned to remove alignment (3 epochs QLoRa). |
|
|
|
### Model Description |
|
|
|
<!-- Provide a longer summary of what this model is. --> |
|
|
|
- **Developed by:** Harambe Research |
|
- **Model type:** llama2 |
|
- **Finetuned from model:** [llama2_xs_460M_experimental](https://huggingface.co/ahxt/llama2_xs_460M_experimental) |
|
|
|
### Out-of-Scope Use |
|
|
|
Don't use this to do bad things. Bad things are bad. |
|
<!-- This section addresses misuse, malicious use, and uses that the model will not work well for. --> |
|
|
|
### Recommendations |
|
|
|
<!-- This section is meant to convey recommendations with respect to the bias, risk, and technical limitations. --> |
|
|
|
Users (both direct and downstream) should be aware of the risks, biases and limitations of the model. |
|
|
|
## How to Get Started with the Model |
|
|
|
https://github.com/oobabooga/text-generation-webui |