File size: 1,609 Bytes
2a48f98 b03d05e 2a48f98 b03d05e 2a48f98 b03d05e 2a48f98 b03d05e 2a48f98 b03d05e 03d2730 144603c d6ecac0 144603c |
1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30 31 32 33 34 35 36 37 38 39 |
---
library_name: transformers
tags: []
---
# SOLAR-10.7b-Instruct-truthy-dpo
![orca-bagel](orca-bagel.png)
This model is a finetune of [macadeliccc/SOLAR-10.7b-Instruct-truthy-dpo](https://huggingface.co/macadeliccc/SOLAR-10.7b-Instruct-dpo)
## Process
1. I finetuned upstageai/Solar-10.7b-Instruct-v0.1 with 1 epoch of Intel/orca_dpo_pairs (12.4k samples)
2. I futher finetuned that model with 3 epochs of jondurbin/truthy-dpo-v0.1 (1.04k samples)
3. This process is experimental and the base model linked above is more tested at this time.
## GGUF
Available [here](https://huggingface.co/macadeliccc/SOLAR-10.7b-Instruct-truthy-dpo-GGUF)
## Evaluations
**Evaluated in 4bit**
| Tasks |Version|Filter|n-shot| Metric |Value | |Stderr|
|-------------|-------|------|-----:|--------|-----:|---|-----:|
|arc_challenge|Yaml |none | 0|acc |0.5853|± |0.0144|
| | |none | 0|acc_norm|0.6126|± |0.0142|
|arc_easy |Yaml |none | 0|acc |0.8077|± |0.0081|
| | |none | 0|acc_norm|0.7715|± |0.0086|
|boolq |Yaml |none | 0|acc |0.8630|± |0.0060|
|hellaswag |Yaml |none | 0|acc |0.6653|± |0.0047|
| | |none | 0|acc_norm|0.8498|± |0.0036|
|openbookqa |Yaml |none | 0|acc |0.3460|± |0.0213|
| | |none | 0|acc_norm|0.4660|± |0.0223|
|piqa |Yaml |none | 0|acc |0.7835|± |0.0096|
| | |none | 0|acc_norm|0.7851|± |0.0096|
|winogrande |Yaml |none | 0|acc |0.7277|± |0.0125|
|