Update README.md
Browse files
README.md
CHANGED
@@ -9,8 +9,8 @@ inference: false
|
|
9 |
## Summary
|
10 |
|
11 |
Databricks’ `dolly-v2-12b`, an instruction-following large language model trained on the Databricks machine learning platform
|
12 |
-
that is licensed for commercial use.
|
13 |
-
[`databricks-dolly-15k`](https://
|
14 |
by Databricks employees in capability domains from the InstructGPT paper, including brainstorming, classification, closed QA, generation,
|
15 |
information extraction, open QA and summarization. `dolly-v2-12b` is not a state-of-the-art model, but does exhibit surprisingly
|
16 |
high quality instruction following behavior not characteristic of the foundation model on which it is based.
|
@@ -20,7 +20,7 @@ high quality instruction following behavior not characteristic of the foundation
|
|
20 |
## Model Overview
|
21 |
`dolly-v2-12b` is a 12 billion parameter causal language model created by [Databricks](https://databricks.com/) that is derived from
|
22 |
[EleutherAI’s](https://www.eleuther.ai/) [Pythia-12b](https://huggingface.co/EleutherAI/pythia-12b) and fine-tuned
|
23 |
-
on a ~15K record instruction corpus generated by Databricks employees and released under a permissive license (CC-BY-SA)
|
24 |
|
25 |
## Usage
|
26 |
|
|
|
9 |
## Summary
|
10 |
|
11 |
Databricks’ `dolly-v2-12b`, an instruction-following large language model trained on the Databricks machine learning platform
|
12 |
+
that is licensed for commercial use. Based on `pythia-12b`, Dolly is trained on ~15k instruction/response fine tuning records
|
13 |
+
[`databricks-dolly-15k`](https://github.com/databrickslabs/dolly/tree/master/data) generated
|
14 |
by Databricks employees in capability domains from the InstructGPT paper, including brainstorming, classification, closed QA, generation,
|
15 |
information extraction, open QA and summarization. `dolly-v2-12b` is not a state-of-the-art model, but does exhibit surprisingly
|
16 |
high quality instruction following behavior not characteristic of the foundation model on which it is based.
|
|
|
20 |
## Model Overview
|
21 |
`dolly-v2-12b` is a 12 billion parameter causal language model created by [Databricks](https://databricks.com/) that is derived from
|
22 |
[EleutherAI’s](https://www.eleuther.ai/) [Pythia-12b](https://huggingface.co/EleutherAI/pythia-12b) and fine-tuned
|
23 |
+
on a [~15K record instruction corpus](https://github.com/databrickslabs/dolly/tree/master/data) generated by Databricks employees and released under a permissive license (CC-BY-SA)
|
24 |
|
25 |
## Usage
|
26 |
|