marcelovidigal commited on
Commit
bf44cea
1 Parent(s): 28e186b

Training in progress, epoch 2

Browse files
model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:e6e816fdda6bc8f7c7116b4bc7cf956f6e9b22468f07ff8a604a912b470a4d5b
3
  size 267832560
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:806bf53cb67d164eceb27c2bda7203b50c1bc8336ad79beead3992c948e3f3e0
3
  size 267832560
wandb/debug-internal.log CHANGED
The diff for this file is too large to render. See raw diff
 
wandb/run-20240924_172630-x9iddikd/files/config.yaml CHANGED
@@ -78,6 +78,18 @@ _wandb:
78
  5: 1
79
  6:
80
  - 1
 
 
 
 
 
 
 
 
 
 
 
 
81
  vocab_size:
82
  desc: null
83
  value: 30522
 
78
  5: 1
79
  6:
80
  - 1
81
+ - 1: train/loss
82
+ 5: 1
83
+ 6:
84
+ - 1
85
+ - 1: train/grad_norm
86
+ 5: 1
87
+ 6:
88
+ - 1
89
+ - 1: train/learning_rate
90
+ 5: 1
91
+ 6:
92
+ - 1
93
  vocab_size:
94
  desc: null
95
  value: 30522
wandb/run-20240924_172630-x9iddikd/files/output.log CHANGED
@@ -19,3 +19,5 @@ dados_tokenizados:
19
  warnings.warn(
20
  Some weights of DistilBertForSequenceClassification were not initialized from the model checkpoint at distilbert/distilbert-base-uncased and are newly initialized: ['classifier.bias', 'classifier.weight', 'pre_classifier.bias', 'pre_classifier.weight']
21
  You should probably TRAIN this model on a down-stream task to be able to use it for predictions and inference.
 
 
 
19
  warnings.warn(
20
  Some weights of DistilBertForSequenceClassification were not initialized from the model checkpoint at distilbert/distilbert-base-uncased and are newly initialized: ['classifier.bias', 'classifier.weight', 'pre_classifier.bias', 'pre_classifier.weight']
21
  You should probably TRAIN this model on a down-stream task to be able to use it for predictions and inference.
22
+ {'eval_loss': 0.21834564208984375, 'eval_accuracy': 0.938, 'eval_runtime': 30.3683, 'eval_samples_per_second': 32.929, 'eval_steps_per_second': 2.075, 'epoch': 1.0}
23
+ {'loss': 0.2031, 'grad_norm': 1.1480563879013062, 'learning_rate': 1.2e-05, 'epoch': 2.0}
wandb/run-20240924_172630-x9iddikd/files/wandb-summary.json CHANGED
@@ -1 +1 @@
1
- {"eval/loss": 0.21834564208984375, "eval/accuracy": 0.938, "eval/runtime": 30.3683, "eval/samples_per_second": 32.929, "eval/steps_per_second": 2.075, "train/epoch": 1.0, "train/global_step": 250, "_timestamp": 1727211087.91596, "_runtime": 1497.0430510044098, "_step": 0}
 
1
+ {"eval/loss": 0.19427122175693512, "eval/accuracy": 0.938, "eval/runtime": 42.2287, "eval/samples_per_second": 23.681, "eval/steps_per_second": 1.492, "train/epoch": 2.0, "train/global_step": 500, "_timestamp": 1727212675.9347782, "_runtime": 3085.0618691444397, "_step": 2, "train/loss": 0.2031, "train/grad_norm": 1.1480563879013062, "train/learning_rate": 1.2e-05}
wandb/run-20240924_172630-x9iddikd/logs/debug-internal.log CHANGED
The diff for this file is too large to render. See raw diff
 
wandb/run-20240924_172630-x9iddikd/run-x9iddikd.wandb CHANGED
Binary files a/wandb/run-20240924_172630-x9iddikd/run-x9iddikd.wandb and b/wandb/run-20240924_172630-x9iddikd/run-x9iddikd.wandb differ