vineetsharma
commited on
Commit
•
a8c6a98
1
Parent(s):
909c13a
Update README.md
Browse files
README.md
CHANGED
@@ -12,7 +12,8 @@ should probably proofread and complete it, then remove this comment. -->
|
|
12 |
|
13 |
# qlora-adapter-Llama-2-7b-hf-databricks-dolly-15k
|
14 |
|
15 |
-
This model is a fine-tuned version of [meta-llama/Llama-2-7b-hf](https://huggingface.co/meta-llama/Llama-2-7b-hf) on
|
|
|
16 |
It achieves the following results on the evaluation set:
|
17 |
- Loss: 1.1313
|
18 |
|
@@ -28,6 +29,8 @@ More information needed
|
|
28 |
|
29 |
More information needed
|
30 |
|
|
|
|
|
31 |
## Training procedure
|
32 |
|
33 |
### Training hyperparameters
|
|
|
12 |
|
13 |
# qlora-adapter-Llama-2-7b-hf-databricks-dolly-15k
|
14 |
|
15 |
+
This model is a fine-tuned version of [meta-llama/Llama-2-7b-hf](https://huggingface.co/meta-llama/Llama-2-7b-hf) on the [databricks/databricks-dolly-15k](https://huggingface.co/datasets/databricks/databricks-dolly-15k) dataset.
|
16 |
+
|
17 |
It achieves the following results on the evaluation set:
|
18 |
- Loss: 1.1313
|
19 |
|
|
|
29 |
|
30 |
More information needed
|
31 |
|
32 |
+
Train on RTX A5000 - 24GB GPU. The training took 3 hours 31 mins on the datasets with 12008 train samples and 1501 validation samples
|
33 |
+
|
34 |
## Training procedure
|
35 |
|
36 |
### Training hyperparameters
|