Update README.md
Browse files
README.md
CHANGED
@@ -9,7 +9,8 @@ model-index:
|
|
9 |
|
10 |
# Llama3_8B_Odia_Unsloth
|
11 |
|
12 |
-
Llama3_8B_Odia_Unsloth is a fine-tuned Odia large language model with 8 billion parameters, and it is based on Llama3. The model is fine-tuned on the 171k Odia instruction set including domain and cultural information.
|
|
|
13 |
|
14 |
For more details about the model, data, training procedure, and evaluations, go through the blog [post]().
|
15 |
|
@@ -27,7 +28,7 @@ If you find this model useful, please consider giving 👏 and citing:
|
|
27 |
```
|
28 |
@misc{Llama3_8B_Odia_Unsloth,
|
29 |
author = {Shantipriya Parida and Sambit Sekhar and Debasish Dhal and Shakshi Panwar},
|
30 |
-
title = {OdiaGenAI
|
31 |
year = {2024},
|
32 |
publisher = {Hugging Face},
|
33 |
journal = {Hugging Face repository},
|
|
|
9 |
|
10 |
# Llama3_8B_Odia_Unsloth
|
11 |
|
12 |
+
Llama3_8B_Odia_Unsloth is a fine-tuned Odia large language model with 8 billion parameters, and it is based on Llama3. The model is fine-tuned on the 171k Odia instruction set including domain and cultural information.
|
13 |
+
The fine-tuning uses Unsloth for faster training.
|
14 |
|
15 |
For more details about the model, data, training procedure, and evaluations, go through the blog [post]().
|
16 |
|
|
|
28 |
```
|
29 |
@misc{Llama3_8B_Odia_Unsloth,
|
30 |
author = {Shantipriya Parida and Sambit Sekhar and Debasish Dhal and Shakshi Panwar},
|
31 |
+
title = {OdiaGenAI Releases Llama3 Fine-tuned Model for the Odia Language},
|
32 |
year = {2024},
|
33 |
publisher = {Hugging Face},
|
34 |
journal = {Hugging Face repository},
|