macadeliccc's picture
Update README.md
46408d1 verified
|
raw
history blame
1.88 kB
metadata
library_name: transformers
tags: []

SOLAR-10.7b-Instruct-truthy-dpo

orca-bagel

This model is a finetune of macadeliccc/SOLAR-10.7b-Instruct-truthy-dpo

Process

  1. I finetuned upstageai/Solar-10.7b-Instruct-v0.1 with 1 epoch of Intel/orca_dpo_pairs (12.4k samples)
  2. I futher finetuned that model with 3 epochs of jondurbin/truthy-dpo-v0.1 (1.04k samples)
  3. This process is experimental and the base model linked above is more tested at this time.

GGUF

Available here

Evaluations

----Benchmark Complete---- + 2024-01-26 20:57:38 + Time taken: 25.4 mins + Prompt Format: ChatML + Model: macadeliccc/SOLAR-10.7b-Instruct-truthy-dpo-GGUF + Score (v2): 74.11 + Parseable: 171.0

Batch completed Time taken: 25.5 mins

Evaluated in 4bit

Tasks Version Filter n-shot Metric Value Stderr
arc_challenge Yaml none 0 acc 0.5853 ± 0.0144
none 0 acc_norm 0.6126 ± 0.0142
arc_easy Yaml none 0 acc 0.8077 ± 0.0081
none 0 acc_norm 0.7715 ± 0.0086
boolq Yaml none 0 acc 0.8630 ± 0.0060
hellaswag Yaml none 0 acc 0.6653 ± 0.0047
none 0 acc_norm 0.8498 ± 0.0036
openbookqa Yaml none 0 acc 0.3460 ± 0.0213
none 0 acc_norm 0.4660 ± 0.0223
piqa Yaml none 0 acc 0.7835 ± 0.0096
none 0 acc_norm 0.7851 ± 0.0096
winogrande Yaml none 0 acc 0.7277 ± 0.0125