Llama3-8B_Traditional_Chinese_roleplay_chat_GGUF

Model description

這是一個於meta-llama/Meta-Llama-3-8B-Instruct上訓練的GGUF,其資料集使用了大量的AI生成的對話範例,使meta-llama/Meta-Llama-3-8B-Instruct能夠盡可能地維持原本充滿活力的對話方式,並能在繁體中文的情況下更加擅長去扮演除了助手以外的角色

This is a GGUF trained on Meta-Llama-3-8B-Instruct, with its dataset consisting of a large number of AI-generated conversation examples. This training allows Meta-Llama-3-8B-Instruct to maintain its original vibrant conversational style as much as possible and excel in roles other than just being an assistant, especially in Chinese contexts.

Training Dataset

微調用的資料集由少量個人撰寫與以此為基礎生成的大量AI生成對話內容組成,使用alpaca-format,約9千條instruction、共12.6MB的大小

The fine-tuning dataset used consists of a small number of personally written conversations and a large amount of AI-generated dialogue content based on these, utilizing the alpaca-format. It comprises approximately 9,000 instructions in total and has a size of 12.6MB.

Training

使用UnslothAI提供的Alpaca + Llama-3 8b Unsloth 2x faster finetuning.ipynb,於Google cloab上調用L4進行微調,訓練參數除將max_steps=60,改為num_train_epochs = 1外其他參數照舊

Using UnsothAI's "Alpaca + Llama-3 8b Unsloth 2x faster finetuning.ipynb" on colab,using L4 GPU to fintuning,only change max_steps=60 to num_train_epochs = 1 in TrainingArguments

Using

這是個GGUF,你需要藉由llama.cpp來使用他

Prompt需要撰寫角色人格、與個性;如有必要則需要另寫對話範例讓AI更好理解如何扮演角色,後續只要再接對話紀錄即可

使用的Prompt範例如下:

小明是一個朝九晚五的上班族,個性樂觀開朗,雖然領的薪水穿不暖又餓不死,但他仍然努力地活在當下
以下是小明與使用者的對話
小明:*今天是你上班的第一天,小明被指派為你的指導員,協助你快速理解公司文化,並讓你能快速上手你的工作,在會議室和你講了數十分鐘的簡報後終於初步介紹完畢*...以上是我們公司大概在做甚麼,現在,你有甚麼問題想要提問的嗎?
使用者:你聽起來有那麼一點倦怠,你有離職的打算嗎?
小明:

Known issues

對話長到一定程度會開始重複回應

License

這個GGUF以CC BY-SA 4.0作為分享

整體使用請遵照Meta的社群許可,不要將其作於非法用途或生成不適當的內容

Downloads last month
10
GGUF
Model size
8.03B params
Architecture
llama

4-bit

Inference Examples
Inference API (serverless) does not yet support GGUF models for this pipeline type.

Dataset used to train Tarklanse/Llama3-8B_Traditional_Chinese_roleplay_chat_GGUF