Starlette!
Update README.md
c3cc9b0 verified
---
library_name: transformers
license: apache-2.0
language:
- en
tags:
- MemGPT
- function
- function calling
---
This is an uncensored release of DPO version of [MemGPT](https://github.com/cpacker/MemGPT) Language Model.
# WARNING
This model is **UNCENSORED**. That means this model is highly compliant to any requests, even unethical and potentially dangerous ones. I do not take any responsibility whatsoever for any damage caused by the model in this repo.
# Model Description
This repository contains an uncensored model of [Mistral 7B Instruct](https://huggingface.co/mistralai/Mistral-7B-Instruct-v0.2). This model is specifically designed for operating within function calling environment in MemGPT. It demonstrates comparable performances to GPT-4 when it comes to working with MemGPT.
# Key Features
* Function calling
* Dedicated to working with MemGPT
* Supports medium-length context, up to sequences of 8,192
# Prompt Format
This model uses **ChatML** prompt format:
```
<|im_start|>system
{system_instruction}<|im_end|>
<|im_start|>user
{user_message}<|im_end|>
<|im_start|>assistant
{assistant_response}<|im_end|>
```
# Usage
This model is designed to be ran on multiple backends, such as [oogabooga's textgen WebUI](https://github.com/oobabooga/text-generation-webui).
Simply install your preferred backend, and then load up this model.
Then, configure MemGPT using `memgpt configure`, and chat with MemGPT via `memgpt run` command!
# Model Details
* Developed by: @starsnatched
* Model type: This repo contains a language model based on the transformer decoder architecture.
* Language: English
* Contact: For any questions, concerns or comments about this model, please contact me at Discord, @starsnatched.
# Training Infrastructure
* Hardware: The model in this repo was trained on 2x A100 80GB GPUs.
# Intended Use
The model is designed to be used as the base model for MemGPT agents.
# Limitations and Risks
The model may exhibit unreliable, unsafe, or biased behaviours. Please double check the results this model may produce.