student-abdullah commited on
Commit
ec26ba7
1 Parent(s): 3aa581a

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +31 -14
README.md CHANGED
@@ -1,22 +1,39 @@
1
  ---
2
- base_model: unsloth/Llama-3.2-1B
3
  language:
4
  - en
5
- license: apache-2.0
6
- tags:
7
- - text-generation-inference
8
- - transformers
9
- - unsloth
10
- - llama
11
- - trl
12
  ---
13
-
14
- # Uploaded model
15
 
16
  - **Developed by:** student-abdullah
17
- - **License:** apache-2.0
18
- - **Finetuned from model :** unsloth/Llama-3.2-1B
19
-
20
- This llama model was trained 2x faster with [Unsloth](https://github.com/unslothai/unsloth) and Huggingface's TRL library.
21
 
 
 
22
  [<img src="https://raw.githubusercontent.com/unslothai/unsloth/main/images/unsloth%20made%20with%20love.png" width="200"/>](https://github.com/unslothai/unsloth)
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
  ---
2
+ base_model: meta-llama/Llama-3.2-1B
3
  language:
4
  - en
5
+ - hi
6
+ datasets:
7
+ - student-abdullah/BigPharma_Generic_Q-A_Format_Augemented_Hinglish_Dataset
 
 
 
 
8
  ---
9
+ # LoRA Adapter Layers!
10
+ # Uploaded model
11
 
12
  - **Developed by:** student-abdullah
13
+ - **Finetuned from model:** meta-llama/Llama-3.2-1B
14
+ - **Created on:** 29th September, 2024
15
+ - **Full model:** student-abdullah/Llama3.2_Medicine-Hinglish-Dataset_Fine-Tuned_29-09
 
16
 
17
+ ---
18
+ # Acknowledgement
19
  [<img src="https://raw.githubusercontent.com/unslothai/unsloth/main/images/unsloth%20made%20with%20love.png" width="200"/>](https://github.com/unslothai/unsloth)
20
+
21
+ ---
22
+ # Model Description
23
+ This LoRA adapter layer model is fine-tuned from the meta-llama/Llama-3.2-1B base model to specialisation related to generic medications under the PMBJP scheme. The fine-tuning process included the following hyperparameters:
24
+
25
+ - Fine Tuning Template: Llama Q&A
26
+ - Max Tokens: 512
27
+ - LoRA Alpha: 32
28
+ - LoRA Rank (r): 128
29
+ - Learning rate: 1.5e-4
30
+ - Gradient Accumulation Steps: 4
31
+ - Batch Size: 8
32
+
33
+ ---
34
+ # Model Quantitative Performace
35
+ - Training Quantitative Loss: 0.1207 (at final 800th epoch)
36
+
37
+ ---
38
+ # Limitations
39
+ - This is not a fully compiled model, rather just LoRA layers