Create README.md
Browse files
README.md
CHANGED
@@ -54,13 +54,8 @@ We have included a [chat template](https://huggingface.co/docs/transformers/main
|
|
54 |
## Intended uses & limitations
|
55 |
|
56 |
The model was initially fine-tuned on a filtered and preprocessed of the [Tulu V2 mix dataset](https://huggingface.co/datasets/allenai/tulu-v2-sft-mixture), which contains a diverse range of human created instructions and synthetic dialogues generated primarily by other LLMs.
|
57 |
-
We then further
|
58 |
-
|
59 |
-
## Bias, Risks, and Limitations
|
60 |
-
|
61 |
-
The Tulu models have not been aligned to generate safe completions within the RLHF phase or deployed with in-the-loop filtering of responses like ChatGPT, so the model can produce problematic outputs (especially when prompted to do so).
|
62 |
-
It is also unknown what the size and composition of the corpus was used to train the base Llama 2 models, however it is likely to have included a mix of Web data and technical sources like books and code. See the [Falcon 180B model card](https://huggingface.co/tiiuae/falcon-180B#training-data) for an example of this.
|
63 |
-
|
64 |
|
65 |
### Training hyperparameters
|
66 |
|
|
|
54 |
## Intended uses & limitations
|
55 |
|
56 |
The model was initially fine-tuned on a filtered and preprocessed of the [Tulu V2 mix dataset](https://huggingface.co/datasets/allenai/tulu-v2-sft-mixture), which contains a diverse range of human created instructions and synthetic dialogues generated primarily by other LLMs.
|
57 |
+
We then further trained the model with a [Jax RM trainer](https://github.com/hamishivi/EasyLM/blob/main/EasyLM/models/llama/llama_train_rm.py) built on [EasyLM](https://github.com/young-geng/EasyLM) on the dataset mentioned above.
|
58 |
+
This model is meant as a research artefact.
|
|
|
|
|
|
|
|
|
|
|
59 |
|
60 |
### Training hyperparameters
|
61 |
|