Update README.md
Browse files
README.md
CHANGED
@@ -53,8 +53,8 @@ Minotaur 13B model is fine-tuned on the following openly available datasets:
|
|
53 |
- custom sysnthetic datasets around misconceptions, in-context qa, jokes, N-tasks problems, and context-insensitivity
|
54 |
- ARC-Easy & ARC-Challenge - instruct augmented for detailed responses, derived from the `train` split
|
55 |
- [hellaswag](https://huggingface.co/datasets/hellaswag) - 30K+ rows of instruct augmented for detailed explanations w 30K+ rows, derived from the `train` split
|
56 |
-
- [riddle_sense](https://huggingface.co/datasets/riddle_sense) - instruct augmented
|
57 |
-
- [gsm8k](https://huggingface.co/datasets/gsm8k) - instruct augmented
|
58 |
- prose generation
|
59 |
|
60 |
# Shoutouts
|
@@ -76,7 +76,7 @@ Minotaur was built with [Axolotl](https://github.com/OpenAccess-AI-Collective/ax
|
|
76 |
|
77 |
## Bias, Risks, and Limitations
|
78 |
Minotaur has not been aligned to human preferences with techniques like RLHF or deployed with in-the-loop filtering of responses like ChatGPT, so the model can produce problematic outputs (especially when prompted to do so).
|
79 |
-
Minotaur was fine-tuned from the base model
|
80 |
|
81 |
## Benchmarks
|
82 |
|
|
|
53 |
- custom sysnthetic datasets around misconceptions, in-context qa, jokes, N-tasks problems, and context-insensitivity
|
54 |
- ARC-Easy & ARC-Challenge - instruct augmented for detailed responses, derived from the `train` split
|
55 |
- [hellaswag](https://huggingface.co/datasets/hellaswag) - 30K+ rows of instruct augmented for detailed explanations w 30K+ rows, derived from the `train` split
|
56 |
+
- [riddle_sense](https://huggingface.co/datasets/riddle_sense) - instruct augmented, derived from the `train` split
|
57 |
+
- [gsm8k](https://huggingface.co/datasets/gsm8k) - instruct augmented, derived from the `train` split
|
58 |
- prose generation
|
59 |
|
60 |
# Shoutouts
|
|
|
76 |
|
77 |
## Bias, Risks, and Limitations
|
78 |
Minotaur has not been aligned to human preferences with techniques like RLHF or deployed with in-the-loop filtering of responses like ChatGPT, so the model can produce problematic outputs (especially when prompted to do so).
|
79 |
+
Minotaur was fine-tuned from the base model LLaMA-13B, please refer to its model card's Limitations Section for relevant information. (included below)
|
80 |
|
81 |
## Benchmarks
|
82 |
|