ChuckMcSneed commited on
Commit
68c75cc
·
verified ·
1 Parent(s): 16527bc

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +23 -0
README.md CHANGED
@@ -16,6 +16,29 @@ datasets:
16
  - Locutusque/function-calling-chatml
17
  - internlm/Agent-FLAN
18
  ---
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
19
 
20
  # Dolphin 2.9.1 DBRX 🐬
21
 
 
16
  - Locutusque/function-calling-chatml
17
  - internlm/Agent-FLAN
18
  ---
19
+ # GGUF fix
20
+ This is the same model as [cognitivecomputations/dolphin-2.9.1-dbrx](https://huggingface.co/cognitivecomputations/dolphin-2.9.1-dbrx), but with [gguf fixes made by Kenjoyer](https://huggingface.co/Kenjoyer/dolphin-2.9.1-dbrx-llamacppfix) applied(thanks a lot!). This model can be converted into gguf using llama.cpp.
21
+
22
+ # Benchmarks and personal opinion
23
+ ### NeoEvalPlusN_benchmark
24
+ [My meme benchmark.](https://huggingface.co/datasets/ChuckMcSneed/NeoEvalPlusN_benchmark)
25
+ |Name |Quant|Size |B |C |D |S |P |total|BCD|SP |
26
+ |---------------------------------------------|-----|------|---|---|---|----|----|-----|---|----|
27
+ |cognitivecomputations/dolphin-2.9.1-dbrx |Q6_K |16x12B|3 |1 |3 |4 |6 |17 |7 |10 |
28
+ |cognitivecomputations/dolphin-2.9.1-qwen-110b|Q6_K |110B |0 |1 |3 |3.75|4.25|12 |4 |8 |
29
+ |databricks/dbrx-instruct |Q6_K |16x12B|0 |0 |0 |6.5 |4.5 |11 |0 |11 |
30
+ |cognitivecomputations/dolphin-2.2-70b |Q6_K |70B |0 |1 |1 |4.5 |4.5 |11 |2 |9 |
31
+ |Maximum |n/a |n/a |3 |2 |3 |8 |6 |22 |8 |14 |
32
+ More compliant than the official instruct tune(BCD). To my surprise, performed much better overall than qwen-110b tuned on the same dataset.
33
+ Wrote 6 perfect poems(P column), which is **very** unusual.
34
+ Only models from goliath family and more recent llama-3-70b-instruct could do that.
35
+ Stylized writing tests(S column) were a bit disappointing, Dolphin is not famous for that.
36
+ In practical use, did perform better than the official tune. Still knows a lot, just like the official tune.
37
+ Writing is not great, wouldn't use it over Command-r+, unless I need to know some obscure facts. Feels like quantization hurts it a lot more than dense models.
38
+
39
+ Verdict: Meh, just like the other dolphins. Eric, no disrespect, but you need to get better datasets. GPTslop really hurts practical performance.
40
+
41
+ # Original model card below
42
 
43
  # Dolphin 2.9.1 DBRX 🐬
44