Locutusque
commited on
Commit
•
14128e0
1
Parent(s):
11130c4
Update README.md
Browse files
README.md
CHANGED
@@ -29,7 +29,7 @@ inference:
|
|
29 |
top_p: 0.34
|
30 |
top_k: 30
|
31 |
max_new_tokens: 250
|
32 |
-
repetition_penalty: 1.
|
33 |
---
|
34 |
# LocutusqueXFelladrin-TinyMistral248M-Instruct
|
35 |
This model was created by merging Locutusque/TinyMistral-248M-Instruct and Felladrin/TinyMistral-248M-SFT-v4 using mergekit. After the two models were merged, the resulting model was further trained on ~20,000 examples on the Locutusque/inst_mix_v2_top_100k at a low learning rate to further normalize weights. The following is the YAML config used to merge:
|
|
|
29 |
top_p: 0.34
|
30 |
top_k: 30
|
31 |
max_new_tokens: 250
|
32 |
+
repetition_penalty: 1.16
|
33 |
---
|
34 |
# LocutusqueXFelladrin-TinyMistral248M-Instruct
|
35 |
This model was created by merging Locutusque/TinyMistral-248M-Instruct and Felladrin/TinyMistral-248M-SFT-v4 using mergekit. After the two models were merged, the resulting model was further trained on ~20,000 examples on the Locutusque/inst_mix_v2_top_100k at a low learning rate to further normalize weights. The following is the YAML config used to merge:
|