File size: 1,609 Bytes
2a48f98
 
 
 
b03d05e
2a48f98
b03d05e
2a48f98
b03d05e
2a48f98
b03d05e
2a48f98
b03d05e
 
03d2730
 
 
 
144603c
 
 
 
d6ecac0
 
144603c
 
 
 
 
 
 
 
 
 
 
 
 
 
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
---
library_name: transformers
tags: []
---
# SOLAR-10.7b-Instruct-truthy-dpo

![orca-bagel](orca-bagel.png)

This model is a finetune of [macadeliccc/SOLAR-10.7b-Instruct-truthy-dpo](https://huggingface.co/macadeliccc/SOLAR-10.7b-Instruct-dpo)

## Process

1. I finetuned upstageai/Solar-10.7b-Instruct-v0.1 with 1 epoch of Intel/orca_dpo_pairs (12.4k samples)
2. I futher finetuned that model with 3 epochs of jondurbin/truthy-dpo-v0.1 (1.04k samples)
3. This process is experimental and the base model linked above is more tested at this time.

## GGUF

Available [here](https://huggingface.co/macadeliccc/SOLAR-10.7b-Instruct-truthy-dpo-GGUF)

## Evaluations

**Evaluated in 4bit**

|    Tasks    |Version|Filter|n-shot| Metric |Value |   |Stderr|
|-------------|-------|------|-----:|--------|-----:|---|-----:|
|arc_challenge|Yaml   |none  |     0|acc     |0.5853|±  |0.0144|
|             |       |none  |     0|acc_norm|0.6126|±  |0.0142|
|arc_easy     |Yaml   |none  |     0|acc     |0.8077|±  |0.0081|
|             |       |none  |     0|acc_norm|0.7715|±  |0.0086|
|boolq        |Yaml   |none  |     0|acc     |0.8630|±  |0.0060|
|hellaswag    |Yaml   |none  |     0|acc     |0.6653|±  |0.0047|
|             |       |none  |     0|acc_norm|0.8498|±  |0.0036|
|openbookqa   |Yaml   |none  |     0|acc     |0.3460|±  |0.0213|
|             |       |none  |     0|acc_norm|0.4660|±  |0.0223|
|piqa         |Yaml   |none  |     0|acc     |0.7835|±  |0.0096|
|             |       |none  |     0|acc_norm|0.7851|±  |0.0096|
|winogrande   |Yaml   |none  |     0|acc     |0.7277|±  |0.0125|