Update README.md
Browse files
README.md
CHANGED
@@ -17,6 +17,7 @@ datasets:
|
|
17 |
- camel-ai/biology
|
18 |
- camel-ai/physics
|
19 |
- camel-ai/chemistry
|
|
|
20 |
|
21 |
inference: false
|
22 |
---
|
@@ -25,7 +26,7 @@ inference: false
|
|
25 |
|
26 |
# Minotaur MPT 7B
|
27 |
|
28 |
-
Minotaur 7B is an instruct fine-tuned model on top of MPT-7B.
|
29 |
|
30 |
Questions, comments, feedback, looking to donate, or want to help? Reach out on our [Discord](https://discord.gg/PugNNHAF5r) or email [wing@openaccessaicollective.org](mailto:wing@openaccessaicollective.org)
|
31 |
|
@@ -36,25 +37,25 @@ Chat only style prompts using `USER:`,`ASSISTANT:`.
|
|
36 |
|
37 |
Minotaur 7B model is fine-tuned on the following datasets:
|
38 |
|
39 |
-
- [riddle_sense](https://huggingface.co/datasets/riddle_sense) - instruct augmented
|
40 |
-
- hellaswag, updated for detailed explanations w 30K+ rows
|
41 |
-
- [gsm8k](https://huggingface.co/datasets/gsm8k) - instruct augmented
|
42 |
- [WizardLM](https://huggingface.co/datasets/ehartford/WizardLM_alpaca_evol_instruct_70k_unfiltered)
|
43 |
- [subset of QingyiSi/Alpaca-CoT for roleplay and CoT](https://huggingface.co/QingyiSi/Alpaca-CoT)
|
44 |
- [GPTeacher-General-Instruct](https://huggingface.co/datasets/teknium/GPTeacher-General-Instruct)
|
45 |
-
- ARC-Easy & ARC-Challenge - instruct augmented for detailed responses, derived from the `train` split
|
46 |
-
- [hellaswag](https://huggingface.co/datasets/hellaswag) - 5K row subset of instruct augmented for concise responses, derived from the `train` split
|
47 |
- [metaeval/ScienceQA_text_only](https://huggingface.co/datasets/metaeval/ScienceQA_text_only) - instruct for concise responses
|
48 |
- [openai/summarize_from_feedback](https://huggingface.co/datasets/openai/summarize_from_feedback) - instruct augmented tl;dr summarization
|
49 |
- [camel-ai/math](https://huggingface.co/datasets/camel-ai/math)
|
50 |
- [camel-ai/physics](https://huggingface.co/datasets/camel-ai/physics)
|
51 |
- [camel-ai/chemistry](https://huggingface.co/datasets/camel-ai/chemistry)
|
52 |
- [camel-ai/biology](https://huggingface.co/datasets/camel-ai/biology)
|
53 |
-
-
|
|
|
|
|
|
|
|
|
|
|
54 |
|
55 |
# Shoutouts
|
56 |
|
57 |
-
Special thanks to Nanobit for helping with Axolotl
|
58 |
|
59 |
# Demo
|
60 |
|
|
|
17 |
- camel-ai/biology
|
18 |
- camel-ai/physics
|
19 |
- camel-ai/chemistry
|
20 |
+
- winglian/evals
|
21 |
|
22 |
inference: false
|
23 |
---
|
|
|
26 |
|
27 |
# Minotaur MPT 7B
|
28 |
|
29 |
+
Minotaur 7B is an instruct fine-tuned model on top of MPT-7B. Minotaur 7B is fine-tuned on only completely open datasets making this model reproducible by anyone.
|
30 |
|
31 |
Questions, comments, feedback, looking to donate, or want to help? Reach out on our [Discord](https://discord.gg/PugNNHAF5r) or email [wing@openaccessaicollective.org](mailto:wing@openaccessaicollective.org)
|
32 |
|
|
|
37 |
|
38 |
Minotaur 7B model is fine-tuned on the following datasets:
|
39 |
|
|
|
|
|
|
|
40 |
- [WizardLM](https://huggingface.co/datasets/ehartford/WizardLM_alpaca_evol_instruct_70k_unfiltered)
|
41 |
- [subset of QingyiSi/Alpaca-CoT for roleplay and CoT](https://huggingface.co/QingyiSi/Alpaca-CoT)
|
42 |
- [GPTeacher-General-Instruct](https://huggingface.co/datasets/teknium/GPTeacher-General-Instruct)
|
|
|
|
|
43 |
- [metaeval/ScienceQA_text_only](https://huggingface.co/datasets/metaeval/ScienceQA_text_only) - instruct for concise responses
|
44 |
- [openai/summarize_from_feedback](https://huggingface.co/datasets/openai/summarize_from_feedback) - instruct augmented tl;dr summarization
|
45 |
- [camel-ai/math](https://huggingface.co/datasets/camel-ai/math)
|
46 |
- [camel-ai/physics](https://huggingface.co/datasets/camel-ai/physics)
|
47 |
- [camel-ai/chemistry](https://huggingface.co/datasets/camel-ai/chemistry)
|
48 |
- [camel-ai/biology](https://huggingface.co/datasets/camel-ai/biology)
|
49 |
+
- [winglian/evals](https://huggingface.co/datasets/winglian/evals)
|
50 |
+
- custom sysnthetic datasets around misconceptions, in-context qa, jokes, N-tasks problems, and context-insensitivity
|
51 |
+
- ARC-Easy & ARC-Challenge - instruct augmented for detailed responses, derived from the `train` split
|
52 |
+
- [hellaswag](https://huggingface.co/datasets/hellaswag) - 30K+ rows of instruct augmented for detailed explanations w 30K+ rows, derived from the `train` split
|
53 |
+
- [riddle_sense](https://huggingface.co/datasets/riddle_sense) - instruct augmented
|
54 |
+
- [gsm8k](https://huggingface.co/datasets/gsm8k) - instruct augmented
|
55 |
|
56 |
# Shoutouts
|
57 |
|
58 |
+
Special thanks to Nanobit for helping with Axolotl and TheBloke for quantizing these models are more accessible to all.
|
59 |
|
60 |
# Demo
|
61 |
|