e-hossam96 commited on
Commit
62c866d
1 Parent(s): 1965f36

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +17 -8
README.md CHANGED
@@ -7,16 +7,18 @@ tags:
7
  model-index:
8
  - name: arabic-nano-gpt-v1
9
  results: []
 
 
 
 
10
  ---
11
 
12
- <!-- This model card has been generated automatically according to the information the Trainer had access to. You
13
- should probably proofread and complete it, then remove this comment. -->
14
 
15
  # arabic-nano-gpt-v1
16
 
17
  This model is a fine-tuned version of [openai-community/gpt2](https://huggingface.co/openai-community/gpt2) on an unknown dataset.
18
- It achieves the following results on the evaluation set:
19
- - Loss: 3.0267
20
 
21
  ## Model description
22
 
@@ -46,9 +48,9 @@ The following hyperparameters were used during training:
46
  - lr_scheduler_warmup_ratio: 0.01
47
  - num_epochs: 24
48
 
49
- ### Training results
50
 
51
- | Training Loss | Epoch | Step | Validation Loss |
52
  |:-------------:|:-------:|:------:|:---------------:|
53
  | 4.1743 | 0.5849 | 5000 | 3.6616 |
54
  | 3.6165 | 1.1698 | 10000 | 3.4256 |
@@ -90,12 +92,19 @@ The following hyperparameters were used during training:
90
  | 3.1775 | 22.2260 | 190000 | 3.0295 |
91
  | 3.1747 | 22.8109 | 195000 | 3.0284 |
92
  | 3.1724 | 23.3957 | 200000 | 3.0273 |
93
- | 3.1757 | 23.9806 | 205000 | 3.0267 |
94
 
 
 
 
 
 
 
 
95
 
96
  ### Framework versions
97
 
98
  - Transformers 4.45.2
99
  - Pytorch 2.5.0
100
  - Datasets 3.0.1
101
- - Tokenizers 0.20.1
 
7
  model-index:
8
  - name: arabic-nano-gpt-v1
9
  results: []
10
+ datasets:
11
+ - wikimedia/wikipedia
12
+ language:
13
+ - ar
14
  ---
15
 
 
 
16
 
17
  # arabic-nano-gpt-v1
18
 
19
  This model is a fine-tuned version of [openai-community/gpt2](https://huggingface.co/openai-community/gpt2) on an unknown dataset.
20
+ It achieves the following results on the held-out test set:
21
+ - Loss: 3.02885
22
 
23
  ## Model description
24
 
 
48
  - lr_scheduler_warmup_ratio: 0.01
49
  - num_epochs: 24
50
 
51
+ <!-- ### Training results -->
52
 
53
+ <!-- | Training Loss | Epoch | Step | Validation Loss |
54
  |:-------------:|:-------:|:------:|:---------------:|
55
  | 4.1743 | 0.5849 | 5000 | 3.6616 |
56
  | 3.6165 | 1.1698 | 10000 | 3.4256 |
 
92
  | 3.1775 | 22.2260 | 190000 | 3.0295 |
93
  | 3.1747 | 22.8109 | 195000 | 3.0284 |
94
  | 3.1724 | 23.3957 | 200000 | 3.0273 |
95
+ | 3.1757 | 23.9806 | 205000 | 3.0267 | -->
96
 
97
+ ### Training Loss
98
+
99
+ ![image/png](https://cdn-uploads.huggingface.co/production/uploads/63ccee86374057a338e03c1e/WIQvnj-VCCBqvsUlJZ1K_.png)
100
+
101
+ ### Validation Loss
102
+
103
+ ![image/png](https://cdn-uploads.huggingface.co/production/uploads/63ccee86374057a338e03c1e/DmTh4sIODlv1wrxXcedxL.png)
104
 
105
  ### Framework versions
106
 
107
  - Transformers 4.45.2
108
  - Pytorch 2.5.0
109
  - Datasets 3.0.1
110
+ - Tokenizers 0.20.1