Edit model card

Instruct_Phi2_Dolly15K

Fine-tuned from phi2,used Dolly15k for the dataset. 90% for training, 10% validation. Trained for 2.0 epochs using QLora. Trained with 1024 context window.

Model Details

  • Trained by: trained by HenryJJ.
  • Model type: Instruct_Phi2_Dolly15K is an auto-regressive language model based on the phi 2 transformer architecture.
  • Language(s): English
  • License for HenryJJ/Instruct_Phi2_Dolly15K: apache-2.0 license

Prompting

Prompt Template With Context

chatml format

<|im_start|>system
{instruction}<|im_end|>
<|im_start|>user
{prompt}<|im_end|>
<|im_start|>assistant

Prompt Template Without Context

<|im_start|>system
{instruction}<|im_end|>
<|im_start|>assistant

Training script:

Fully opensourced at: https://github.com/hengjiUSTC/learn-llm/blob/main/trl_finetune.py. Run on 1 A10G instance for 4 hours.

python3 trl_finetune.py --config configs/phi2-dolly.yml
Downloads last month
10
Safetensors
Model size
2.78B params
Tensor type
BF16
·
Inference Examples
This model does not have enough activity to be deployed to Inference API (serverless) yet. Increase its social visibility and check back later, or deploy to Inference Endpoints (dedicated) instead.

Dataset used to train HenryJJ/Instruct_Phi2_Dolly15K