DazMashaly commited on
Commit
918177f
1 Parent(s): 2b7532b

Training in progress, epoch 5

Browse files
README.md CHANGED
@@ -1,27 +1,26 @@
1
  ---
2
- base_model: DazMashaly/output
3
  tags:
4
  - generated_from_trainer
5
  model-index:
6
- - name: new_downloads
7
  results: []
8
  ---
9
 
10
  <!-- This model card has been generated automatically according to the information the Trainer had access to. You
11
  should probably proofread and complete it, then remove this comment. -->
12
 
13
- # new_downloads
14
 
15
- This model is a fine-tuned version of [DazMashaly/output](https://huggingface.co/DazMashaly/output) on an unknown dataset.
16
  It achieves the following results on the evaluation set:
17
- - eval_loss: 5.3543
18
  - eval_wer: 1.0
19
  - eval_cer: 1.0
20
- - eval_runtime: 973.8534
21
- - eval_samples_per_second: 2.258
22
- - eval_steps_per_second: 0.018
23
- - epoch: 4.0
24
- - step: 1636
25
 
26
  ## Model description
27
 
@@ -41,7 +40,7 @@ More information needed
41
 
42
  The following hyperparameters were used during training:
43
  - learning_rate: 1e-05
44
- - train_batch_size: 128
45
  - eval_batch_size: 128
46
  - seed: 42
47
  - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
@@ -54,5 +53,5 @@ The following hyperparameters were used during training:
54
 
55
  - Transformers 4.41.2
56
  - Pytorch 2.1.2
57
- - Datasets 2.19.2
58
  - Tokenizers 0.19.1
 
1
  ---
 
2
  tags:
3
  - generated_from_trainer
4
  model-index:
5
+ - name: output
6
  results: []
7
  ---
8
 
9
  <!-- This model card has been generated automatically according to the information the Trainer had access to. You
10
  should probably proofread and complete it, then remove this comment. -->
11
 
12
+ # output
13
 
14
+ This model was trained from scratch on an unknown dataset.
15
  It achieves the following results on the evaluation set:
16
+ - eval_loss: 5.5136
17
  - eval_wer: 1.0
18
  - eval_cer: 1.0
19
+ - eval_runtime: 702.4432
20
+ - eval_samples_per_second: 3.131
21
+ - eval_steps_per_second: 0.026
22
+ - epoch: 2.0
23
+ - step: 818
24
 
25
  ## Model description
26
 
 
40
 
41
  The following hyperparameters were used during training:
42
  - learning_rate: 1e-05
43
+ - train_batch_size: 124
44
  - eval_batch_size: 128
45
  - seed: 42
46
  - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
 
53
 
54
  - Transformers 4.41.2
55
  - Pytorch 2.1.2
56
+ - Datasets 2.20.0
57
  - Tokenizers 0.19.1
config.json CHANGED
@@ -1,5 +1,5 @@
1
  {
2
- "_name_or_path": "DazMashaly/output",
3
  "activation_dropout": 0.0,
4
  "activation_function": "gelu",
5
  "apply_spec_augment": false,
 
1
  {
2
+ "_name_or_path": "DazMashaly/new_downloads",
3
  "activation_dropout": 0.0,
4
  "activation_function": "gelu",
5
  "apply_spec_augment": false,
model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:4ca4e27d8a2950ea1358f16e3c135a63cabeaa10ec29e03c83c8c073803cd5f8
3
  size 290403936
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:fc48c0adcc1e3b3c0bd007518f406e2619082456c1687e6f499e09a28926a2b2
3
  size 290403936
runs/Jun25_06-08-05_557d2a8f2caf/events.out.tfevents.1719295686.557d2a8f2caf.34.0 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:e0db750645d9eff514f6bc22611b955c656f7723607f3b410db2b63d04ac526d
3
+ size 10374
training_args.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:4df180bbc58c0a7714e8beb26c654c1c1a085db2c02893349f4f27f5b8880fbe
3
  size 5304
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:f73b80aec597d06b3d2e1eea4b8902092e7ee47b50c6ace8e729e22ced3aeffe
3
  size 5304