IggoOnCode's picture
Update README.md
7399485 verified
metadata
'[object Object]': null

Model Card for mamba-2.8b-slimpj-OpenOrca_1ep

This is a finetune of mamba-2.8b-slimpj for instruction following using the OpenOrca dataset.

Model Details

Model Description

This is a finetune of the mamba reference model mamba-2.8b-slimpj from the paper https://arxiv.org/abs/2312.00752

It has been fine-tuned for instruction following using the OpenOrca dataset and training for 1 epoch.

Uses

This model is intended to evaluate fine-tuning results on mamba models.

Usage

Prompt structure

The prompt structure used in fine-tuning is alpaca format:

"### Human:\n%question%\n\n### AI response:\n%response%"

Training Details

Training Data

https://huggingface.co/datasets/Open-Orca/OpenOrca

Training Procedure

Trained using text-generation-webui with code from the mamba_ssm pull request.

Training Hyperparameters

  • Training regime: Trained in bfloat16 with the following parameters:
{
  "trained_model_name": "mamba-2.8b-slimpj-OpenOrc_1ep",
  "save_steps": 500000.0,
  "micro_batch_size": 4,
  "batch_size": 128,
  "epochs": 1.0,
  "learning_rate": "3e-4",
  "lr_scheduler_type": "linear",
  "cutoff_len": 256,
  "dataset": "OpenOrca",
  "eval_dataset": "None",
  "format": "openorca-format",
  "warmup_steps": 100.0,
  "optimizer": "paged_adamw_8bit",
  "hard_cut_string": "\\n\\n\\n",
  "add_eos_token": false,
  "min_chars": 0.0,
}

Reported train_loss was 0.6762700151924311

Results

lm-evaluation-harness results for final model

mamba_ssm (pretrained=mamba-2.8b-slimpj-OpenOrca), gen_kwargs: (None), limit: None, num_fewshot: None, batch_size: auto (32)

Tasks Version Filter n-shot Metric Value Stderr
arc_challenge 1 none 0 acc 0.2594 ± 0.0128
none 0 acc_norm 0.2935 ± 0.0133
arc_easy 1 none 0 acc 0.4390 ± 0.0102
none 0 acc_norm 0.4032 ± 0.0101
boolq 2 none 0 acc 0.5801 ± 0.0086
lambada_openai 1 none 0 perplexity 27.8582 ± 1.1183
none 0 acc 0.3683 ± 0.0067
openbookqa 1 none 0 acc 0.2500 ± 0.0194
none 0 acc_norm 0.3700 ± 0.0216
piqa 1 none 0 acc 0.6817 ± 0.0109
none 0 acc_norm 0.6839 ± 0.0108
winogrande 1 none 0 acc 0.5770 ± 0.0139

lm-evaluation-harness results after half epoch

mamba_ssm (pretrained=mamba-2.8b-slimpj-OpenOrca_1ep-checkpoints/checkpoint-500000), gen_kwargs: (None), limit: None, num_fewshot: None, batch_size: auto (32)

Tasks Version Filter n-shot Metric Value Stderr
arc_challenge 1 none 0 acc 0.2602 ± 0.0128
none 0 acc_norm 0.2833 ± 0.0132
arc_easy 1 none 0 acc 0.4533 ± 0.0102
none 0 acc_norm 0.4125 ± 0.0101
boolq 2 none 0 acc 0.4095 ± 0.0086
lambada_openai 1 none 0 perplexity 30.4832 ± 1.2403
none 0 acc 0.3551 ± 0.0067
openbookqa 1 none 0 acc 0.2420 ± 0.0192
none 0 acc_norm 0.3640 ± 0.0215
piqa 1 none 0 acc 0.6812 ± 0.0109
none 0 acc_norm 0.6730 ± 0.0109
winogrande 1 none 0 acc 0.5588 ± 0.0140

Reference lm-evaluation-harness results for the base model mamba-2.8b-slimpj without fine-tuning

mamba_ssm (pretrained=mamba-2.8b-slimpj), gen_kwargs: (None), limit: None, num_fewshot: None, batch_size: auto (32)

Tasks Version Filter n-shot Metric Value Stderr
arc_challenge 1 none 0 acc 0.3882 ± 0.0142
none 0 acc_norm 0.4155 ± 0.0144
arc_easy 1 none 0 acc 0.7264 ± 0.0091
none 0 acc_norm 0.6814 ± 0.0096
boolq 2 none 0 acc 0.7107 ± 0.0079
lambada_openai 1 none 0 perplexity 5.8770 ± 0.1881
none 0 acc 0.6427 ± 0.0067
openbookqa 1 none 0 acc 0.2860 ± 0.0202
none 0 acc_norm 0.3980 ± 0.0219
piqa 1 none 0 acc 0.7709 ± 0.0098
none 0 acc_norm 0.7813 ± 0.0096
winogrande 1 none 0 acc 0.6614 ± 0.0133

Summary

The models measured perplexity and accuracy got worse, but it's known that that can be an effect of fine-tuning. Perplexity and accuracy improved in the second half of the training, so it's likely that the inital worsening was caused by forcing a prompt structure onto the base model, which was trained only on unstructured text.

The answer quality as percieved by users is yet to be evaluated.

Environmental Impact

  • Hardware Type: RTX 3090
  • Hours used: 118