mpt-30b-v2 / README.md
manojpreveen's picture
Duplicate from iamplus/mpt-30b-v2
bf8efe1 verified
|
raw
history blame contribute delete
No virus
406 Bytes
---
datasets:
- ehartford/dolphin
license: apache-2.0
---
**Base Model :** mosaicml/mpt-30b
**Tool :** MosaicML's llm-foundry (https://github.com/mosaicml/llm-foundry)
**Dataset :** Entire flan3m-GPT3.5 dataset.
**Config yaml with Model Params :** https://huggingface.co/iamplus/mpt-30b-v2/blob/main/mpt-30b_orca.yaml
**Prompt Format :**
```
<system>: [system prompt]
<human>: [question]
<bot>:
```