edumunozsala
commited on
Commit
•
d6f82aa
1
Parent(s):
2b9d925
Upload README.md
Browse files
README.md
CHANGED
@@ -16,7 +16,7 @@ pipeline_tag: text-generation
|
|
16 |
---
|
17 |
|
18 |
|
19 |
-
# LlaMa 2 7b 4-bit Python Coder 👩💻
|
20 |
|
21 |
**LlaMa-2 7b** fine-tuned on the **CodeAlpaca 20k instructions dataset** by using the method **QLoRA** in 4-bit with [PEFT](https://github.com/huggingface/peft) library.
|
22 |
|
@@ -76,6 +76,30 @@ The following `bitsandbytes` quantization config was used during training:
|
|
76 |
### Framework versions
|
77 |
- PEFT 0.4.0
|
78 |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
79 |
### Example of usage
|
80 |
```py
|
81 |
import torch
|
@@ -117,10 +141,10 @@ print(f"Generated instruction:\n{tokenizer.batch_decode(outputs.detach().cpu().n
|
|
117 |
```
|
118 |
@misc {edumunozsala_2023,
|
119 |
author = { {Eduardo Muñoz} },
|
120 |
-
title = { llama-2-7b-int4-python-coder
|
121 |
year = 2023,
|
122 |
url = { https://huggingface.co/edumunozsala/llama-2-7b-int4-python-18k-alpaca },
|
123 |
-
doi = {
|
124 |
publisher = { Hugging Face }
|
125 |
}
|
126 |
```
|
|
|
16 |
---
|
17 |
|
18 |
|
19 |
+
# LlaMa 2 7b 4-bit Python Coder 👩💻 :man_technologist:
|
20 |
|
21 |
**LlaMa-2 7b** fine-tuned on the **CodeAlpaca 20k instructions dataset** by using the method **QLoRA** in 4-bit with [PEFT](https://github.com/huggingface/peft) library.
|
22 |
|
|
|
76 |
### Framework versions
|
77 |
- PEFT 0.4.0
|
78 |
|
79 |
+
### Training metrics
|
80 |
+
{'loss': 1.044, 'learning_rate': 3.571428571428572e-05, 'epoch': 0.01}
|
81 |
+
{'loss': 0.8413, 'learning_rate': 7.142857142857143e-05, 'epoch': 0.01}
|
82 |
+
{'loss': 0.7299, 'learning_rate': 0.00010714285714285715, 'epoch': 0.02}
|
83 |
+
{'loss': 0.6593, 'learning_rate': 0.00014285714285714287, 'epoch': 0.02}
|
84 |
+
{'loss': 0.6309, 'learning_rate': 0.0001785714285714286, 'epoch': 0.03}
|
85 |
+
{'loss': 0.5916, 'learning_rate': 0.00019999757708974043, 'epoch': 0.03}
|
86 |
+
{'loss': 0.5861, 'learning_rate': 0.00019997032069768138, 'epoch': 0.04}
|
87 |
+
{'loss': 0.6118, 'learning_rate': 0.0001999127875580558, 'epoch': 0.04}
|
88 |
+
{'loss': 0.5928, 'learning_rate': 0.00019982499509519857, 'epoch': 0.05}
|
89 |
+
{'loss': 0.5978, 'learning_rate': 0.00019970696989770335, 'epoch': 0.05}
|
90 |
+
{'loss': 0.5791, 'learning_rate': 0.0001995587477103701, 'epoch': 0.06}
|
91 |
+
{'loss': 0.6054, 'learning_rate': 0.00019938037342337933, 'epoch': 0.06}
|
92 |
+
{'loss': 0.5864, 'learning_rate': 0.00019917190105869708, 'epoch': 0.07}
|
93 |
+
{'loss': 0.6159, 'learning_rate': 0.0001989333937537136, 'epoch': 0.08}
|
94 |
+
{'loss': 0.583, 'learning_rate': 0.00019866492374212205, 'epoch': 0.08}
|
95 |
+
{'loss': 0.6066, 'learning_rate': 0.00019836657233204182, 'epoch': 0.09}
|
96 |
+
{'loss': 0.5934, 'learning_rate': 0.00019803842988139374, 'epoch': 0.09}
|
97 |
+
{'loss': 0.5836, 'learning_rate': 0.00019768059577053473, 'epoch': 0.1}
|
98 |
+
{'loss': 0.6021, 'learning_rate': 0.00019729317837215943, 'epoch': 0.1}
|
99 |
+
{'loss': 0.5659, 'learning_rate': 0.00019687629501847898, 'epoch': 0.11}
|
100 |
+
{'loss': 0.5754, 'learning_rate': 0.00019643007196568606, 'epoch': 0.11}
|
101 |
+
{'loss': 0.5936, 'learning_rate': 0.000195954644355717, 'epoch': 0.12}
|
102 |
+
|
103 |
### Example of usage
|
104 |
```py
|
105 |
import torch
|
|
|
141 |
```
|
142 |
@misc {edumunozsala_2023,
|
143 |
author = { {Eduardo Muñoz} },
|
144 |
+
title = { llama-2-7b-int4-python-coder },
|
145 |
year = 2023,
|
146 |
url = { https://huggingface.co/edumunozsala/llama-2-7b-int4-python-18k-alpaca },
|
147 |
+
doi = { },
|
148 |
publisher = { Hugging Face }
|
149 |
}
|
150 |
```
|