winglian commited on
Commit
7381abe
1 Parent(s): 98482f9

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +3 -3
README.md CHANGED
@@ -53,8 +53,8 @@ Minotaur 13B model is fine-tuned on the following openly available datasets:
53
  - custom sysnthetic datasets around misconceptions, in-context qa, jokes, N-tasks problems, and context-insensitivity
54
  - ARC-Easy & ARC-Challenge - instruct augmented for detailed responses, derived from the `train` split
55
  - [hellaswag](https://huggingface.co/datasets/hellaswag) - 30K+ rows of instruct augmented for detailed explanations w 30K+ rows, derived from the `train` split
56
- - [riddle_sense](https://huggingface.co/datasets/riddle_sense) - instruct augmented
57
- - [gsm8k](https://huggingface.co/datasets/gsm8k) - instruct augmented
58
  - prose generation
59
 
60
  # Shoutouts
@@ -76,7 +76,7 @@ Minotaur was built with [Axolotl](https://github.com/OpenAccess-AI-Collective/ax
76
 
77
  ## Bias, Risks, and Limitations
78
  Minotaur has not been aligned to human preferences with techniques like RLHF or deployed with in-the-loop filtering of responses like ChatGPT, so the model can produce problematic outputs (especially when prompted to do so).
79
- Minotaur was fine-tuned from the base model MPT-7B, please refer to its model card's Limitations Section for relevant information. (included below)
80
 
81
  ## Benchmarks
82
 
 
53
  - custom sysnthetic datasets around misconceptions, in-context qa, jokes, N-tasks problems, and context-insensitivity
54
  - ARC-Easy & ARC-Challenge - instruct augmented for detailed responses, derived from the `train` split
55
  - [hellaswag](https://huggingface.co/datasets/hellaswag) - 30K+ rows of instruct augmented for detailed explanations w 30K+ rows, derived from the `train` split
56
+ - [riddle_sense](https://huggingface.co/datasets/riddle_sense) - instruct augmented, derived from the `train` split
57
+ - [gsm8k](https://huggingface.co/datasets/gsm8k) - instruct augmented, derived from the `train` split
58
  - prose generation
59
 
60
  # Shoutouts
 
76
 
77
  ## Bias, Risks, and Limitations
78
  Minotaur has not been aligned to human preferences with techniques like RLHF or deployed with in-the-loop filtering of responses like ChatGPT, so the model can produce problematic outputs (especially when prompted to do so).
79
+ Minotaur was fine-tuned from the base model LLaMA-13B, please refer to its model card's Limitations Section for relevant information. (included below)
80
 
81
  ## Benchmarks
82