eluzhnica commited on
Commit
22fc52e
1 Parent(s): 5d43aa9

Modify readme

Browse files
Files changed (1) hide show
  1. README.md +6 -1
README.md CHANGED
@@ -19,6 +19,11 @@ inference: false
19
 
20
  # MPT-7B-Instruct-8k
21
 
 
 
 
 
 
22
  MPT-7B-Instruct-8k is a model for long-form instruction following, especially question-answering on and summarization of longer documents.
23
  It is built by finetuning [MPT-7B-8k](https://huggingface.co/mosaicml/mpt-7b-8k) on [Dolly HHRLHF](https://huggingface.co/datasets/mosaicml/dolly_hhrlhf) derived from the [Databricks Dolly-15k](https://huggingface.co/datasets/databricks/databricks-dolly-15k) and the [Anthropic Helpful and Harmless (HH-RLHF)](https://huggingface.co/datasets/Anthropic/hh-rlhf) datasets. It is also trained on [Competition Math](https://huggingface.co/datasets/competition_math), [Duorc](https://huggingface.co/datasets/duorc), [CoT GSM8k](https://huggingface.co/datasets/conceptofmind/cot_submix_original), [Qasper](https://huggingface.co/datasets/allenai/qasper), [Quality](https://huggingface.co/datasets/emozilla/quality), [Summ Screen FD](https://huggingface.co/datasets/tau/scrolls) and [Spider](https://huggingface.co/datasets/spider).
24
  This is the same dataset that [MPT-30B-Instruct](https://huggingface.co/mosaicml/mpt-30b-instruct) was trained on.
@@ -195,4 +200,4 @@ for open-source foundation models},
195
  note = {Accessed: 2023-06-22},
196
  urldate = {2023-06-22}
197
  }
198
- ```
 
19
 
20
  # MPT-7B-Instruct-8k
21
 
22
+ MPT-7B-Instruct-8K but with gradient checkpointing making it easy to train with LoRA/QLoRA. Not tested yet
23
+
24
+
25
+ Original card below:
26
+
27
  MPT-7B-Instruct-8k is a model for long-form instruction following, especially question-answering on and summarization of longer documents.
28
  It is built by finetuning [MPT-7B-8k](https://huggingface.co/mosaicml/mpt-7b-8k) on [Dolly HHRLHF](https://huggingface.co/datasets/mosaicml/dolly_hhrlhf) derived from the [Databricks Dolly-15k](https://huggingface.co/datasets/databricks/databricks-dolly-15k) and the [Anthropic Helpful and Harmless (HH-RLHF)](https://huggingface.co/datasets/Anthropic/hh-rlhf) datasets. It is also trained on [Competition Math](https://huggingface.co/datasets/competition_math), [Duorc](https://huggingface.co/datasets/duorc), [CoT GSM8k](https://huggingface.co/datasets/conceptofmind/cot_submix_original), [Qasper](https://huggingface.co/datasets/allenai/qasper), [Quality](https://huggingface.co/datasets/emozilla/quality), [Summ Screen FD](https://huggingface.co/datasets/tau/scrolls) and [Spider](https://huggingface.co/datasets/spider).
29
  This is the same dataset that [MPT-30B-Instruct](https://huggingface.co/mosaicml/mpt-30b-instruct) was trained on.
 
200
  note = {Accessed: 2023-06-22},
201
  urldate = {2023-06-22}
202
  }
203
+ ```