MaRyAm1295 commited on
Commit
16a1d88
1 Parent(s): 20d7ec4

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +38 -22
README.md CHANGED
@@ -3,16 +3,23 @@ base_model: meta-llama/Llama-3.1-8B-Instruct
3
  library_name: transformers
4
  model_name: Llama-3.1-8B-KAM
5
  tags:
6
- - generated_from_trainer
7
  - trl
 
8
  - sft
 
9
  licence: license
10
  ---
11
 
 
 
 
12
  # Model Card for Llama-3.1-8B-KAM
13
 
14
- This model is a fine-tuned version of [meta-llama/Llama-3.1-8B-Instruct](https://huggingface.co/meta-llama/Llama-3.1-8B-Instruct).
15
- It has been trained using [TRL](https://github.com/huggingface/trl).
 
 
 
16
 
17
  ## Quick start
18
 
@@ -27,9 +34,35 @@ print(output["generated_text"])
27
 
28
  ## Training procedure
29
 
 
 
 
30
 
 
 
 
 
 
 
 
 
 
 
 
 
31
 
32
- This model was trained with SFT.
 
 
 
 
 
 
 
 
 
 
 
33
 
34
  ### Framework versions
35
 
@@ -37,21 +70,4 @@ This model was trained with SFT.
37
  - Transformers: 4.46.2
38
  - Pytorch: 2.4.0
39
  - Datasets: 3.0.1
40
- - Tokenizers: 0.20.0
41
-
42
- ## Citations
43
-
44
-
45
-
46
- Cite TRL as:
47
-
48
- ```bibtex
49
- @misc{vonwerra2022trl,
50
- title = {{TRL: Transformer Reinforcement Learning}},
51
- author = {Leandro von Werra and Younes Belkada and Lewis Tunstall and Edward Beeching and Tristan Thrush and Nathan Lambert and Shengyi Huang and Kashif Rasul and Quentin Gallouédec},
52
- year = 2020,
53
- journal = {GitHub repository},
54
- publisher = {GitHub},
55
- howpublished = {\url{https://github.com/huggingface/trl}}
56
- }
57
- ```
 
3
  library_name: transformers
4
  model_name: Llama-3.1-8B-KAM
5
  tags:
 
6
  - trl
7
+ - Llama
8
  - sft
9
+ - generated_from_trainer
10
  licence: license
11
  ---
12
 
13
+ <!-- This model card has been generated automatically according to the information the Trainer had access to. You
14
+ should probably proofread and complete it, then remove this comment. -->
15
+
16
  # Model Card for Llama-3.1-8B-KAM
17
 
18
+ This model is a fine-tuned version of [meta-llama/Llama-3.1-8B-Instruct](https://huggingface.co/meta-llama/Llama-3.1-8B-Instruct) on the None dataset.
19
+
20
+ ## Model description
21
+
22
+ More information needed
23
 
24
  ## Quick start
25
 
 
34
 
35
  ## Training procedure
36
 
37
+ This model was trained with SFT.
38
+
39
+ ### Training hyperparameters
40
 
41
+ The following hyperparameters were used during training:
42
+ - learning_rate: 0.0002
43
+ - train_batch_size: 1
44
+ - eval_batch_size: 8
45
+ - seed: 3407
46
+ - gradient_accumulation_steps: 16
47
+ - total_train_batch_size: 8
48
+ - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
49
+ - lr_scheduler_type: linear
50
+ - lr_scheduler_warmup_steps: 20
51
+ - training_steps: 500
52
+ - mixed_precision_training: Native AMP
53
 
54
+ ### Training results
55
+ #### Step &nbsp;&nbsp; Training Loss
56
+ - 50 &nbsp;&nbsp;&nbsp; 2.158200
57
+ - 100 &nbsp;&nbsp;&nbsp; 1.845900
58
+ - 150 &nbsp;&nbsp;&nbsp; 1.832200
59
+ - 200 &nbsp;&nbsp;&nbsp; 1.805300
60
+ - 250 &nbsp;&nbsp;&nbsp; 1.783800
61
+ - 300 &nbsp;&nbsp;&nbsp; 1.767500
62
+ - 350 &nbsp;&nbsp;&nbsp; 1.744800
63
+ - 400 &nbsp;&nbsp;&nbsp; 1.745600
64
+ - 450 &nbsp;&nbsp;&nbsp; 1.749500
65
+ - 500 &nbsp;&nbsp;&nbsp; 1.756100
66
 
67
  ### Framework versions
68
 
 
70
  - Transformers: 4.46.2
71
  - Pytorch: 2.4.0
72
  - Datasets: 3.0.1
73
+ - Tokenizers: 0.20.0