teknium commited on
Commit
2eb41a4
1 Parent(s): 51b7f98

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +4 -10
README.md CHANGED
@@ -38,7 +38,7 @@ This is the SFT + DPO version of Mixtral Hermes 2, we will also be providing an
38
  - GPT4All
39
  - AGIEval
40
  - BigBench
41
- - TruthfulQA
42
  3. [Prompt Format](#prompt-format)
43
  4. [Inference Example Code](#inference-code)
44
  5. [Quantized Models](#quantized-models)
@@ -131,14 +131,6 @@ BigBench:
131
  ```
132
  Average: 49.70
133
 
134
- TruthfulQA:
135
- ```
136
- | Task |Version|Metric|Value | |Stderr|
137
- |-------------|------:|------|-----:|---|-----:|
138
- |truthfulqa_mc| 1|mc1 |0.4162|± |0.0173|
139
- | | |mc2 |0.5783|± |0.0151|
140
- ```
141
-
142
  ## GPT4All
143
 
144
 
@@ -148,9 +140,11 @@ TruthfulQA:
148
  ## BigBench Reasoning Test
149
 
150
 
151
- ## TruthfulQA:
152
 
 
153
 
 
154
 
155
  # Prompt Format
156
 
 
38
  - GPT4All
39
  - AGIEval
40
  - BigBench
41
+ - Comparison to Mixtral-Instruct
42
  3. [Prompt Format](#prompt-format)
43
  4. [Inference Example Code](#inference-code)
44
  5. [Quantized Models](#quantized-models)
 
131
  ```
132
  Average: 49.70
133
 
 
 
 
 
 
 
 
 
134
  ## GPT4All
135
 
136
 
 
140
  ## BigBench Reasoning Test
141
 
142
 
143
+ ## Comparison to Mixtral Instruct:
144
 
145
+ Our benchmarks show gains in many benchmarks against Mixtral Instruct v0.1, on average, beating the flagship Mixtral model.
146
 
147
+ ![image/png](https://cdn-uploads.huggingface.co/production/uploads/6317aade83d8d2fd903192d9/TuB0kC6rLmCCkiGLKB2_j.png)
148
 
149
  # Prompt Format
150