Modify readme
Browse files
README.md
CHANGED
@@ -19,6 +19,11 @@ inference: false
|
|
19 |
|
20 |
# MPT-7B-Instruct-8k
|
21 |
|
|
|
|
|
|
|
|
|
|
|
22 |
MPT-7B-Instruct-8k is a model for long-form instruction following, especially question-answering on and summarization of longer documents.
|
23 |
It is built by finetuning [MPT-7B-8k](https://huggingface.co/mosaicml/mpt-7b-8k) on [Dolly HHRLHF](https://huggingface.co/datasets/mosaicml/dolly_hhrlhf) derived from the [Databricks Dolly-15k](https://huggingface.co/datasets/databricks/databricks-dolly-15k) and the [Anthropic Helpful and Harmless (HH-RLHF)](https://huggingface.co/datasets/Anthropic/hh-rlhf) datasets. It is also trained on [Competition Math](https://huggingface.co/datasets/competition_math), [Duorc](https://huggingface.co/datasets/duorc), [CoT GSM8k](https://huggingface.co/datasets/conceptofmind/cot_submix_original), [Qasper](https://huggingface.co/datasets/allenai/qasper), [Quality](https://huggingface.co/datasets/emozilla/quality), [Summ Screen FD](https://huggingface.co/datasets/tau/scrolls) and [Spider](https://huggingface.co/datasets/spider).
|
24 |
This is the same dataset that [MPT-30B-Instruct](https://huggingface.co/mosaicml/mpt-30b-instruct) was trained on.
|
@@ -195,4 +200,4 @@ for open-source foundation models},
|
|
195 |
note = {Accessed: 2023-06-22},
|
196 |
urldate = {2023-06-22}
|
197 |
}
|
198 |
-
```
|
|
|
19 |
|
20 |
# MPT-7B-Instruct-8k
|
21 |
|
22 |
+
MPT-7B-Instruct-8K but with gradient checkpointing making it easy to train with LoRA/QLoRA. Not tested yet
|
23 |
+
|
24 |
+
|
25 |
+
Original card below:
|
26 |
+
|
27 |
MPT-7B-Instruct-8k is a model for long-form instruction following, especially question-answering on and summarization of longer documents.
|
28 |
It is built by finetuning [MPT-7B-8k](https://huggingface.co/mosaicml/mpt-7b-8k) on [Dolly HHRLHF](https://huggingface.co/datasets/mosaicml/dolly_hhrlhf) derived from the [Databricks Dolly-15k](https://huggingface.co/datasets/databricks/databricks-dolly-15k) and the [Anthropic Helpful and Harmless (HH-RLHF)](https://huggingface.co/datasets/Anthropic/hh-rlhf) datasets. It is also trained on [Competition Math](https://huggingface.co/datasets/competition_math), [Duorc](https://huggingface.co/datasets/duorc), [CoT GSM8k](https://huggingface.co/datasets/conceptofmind/cot_submix_original), [Qasper](https://huggingface.co/datasets/allenai/qasper), [Quality](https://huggingface.co/datasets/emozilla/quality), [Summ Screen FD](https://huggingface.co/datasets/tau/scrolls) and [Spider](https://huggingface.co/datasets/spider).
|
29 |
This is the same dataset that [MPT-30B-Instruct](https://huggingface.co/mosaicml/mpt-30b-instruct) was trained on.
|
|
|
200 |
note = {Accessed: 2023-06-22},
|
201 |
urldate = {2023-06-22}
|
202 |
}
|
203 |
+
```
|