marcelovidigal
commited on
Commit
•
bf44cea
1
Parent(s):
28e186b
Training in progress, epoch 2
Browse files- model.safetensors +1 -1
- wandb/debug-internal.log +0 -0
- wandb/run-20240924_172630-x9iddikd/files/config.yaml +12 -0
- wandb/run-20240924_172630-x9iddikd/files/output.log +2 -0
- wandb/run-20240924_172630-x9iddikd/files/wandb-summary.json +1 -1
- wandb/run-20240924_172630-x9iddikd/logs/debug-internal.log +0 -0
- wandb/run-20240924_172630-x9iddikd/run-x9iddikd.wandb +0 -0
model.safetensors
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
size 267832560
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:806bf53cb67d164eceb27c2bda7203b50c1bc8336ad79beead3992c948e3f3e0
|
3 |
size 267832560
|
wandb/debug-internal.log
CHANGED
The diff for this file is too large to render.
See raw diff
|
|
wandb/run-20240924_172630-x9iddikd/files/config.yaml
CHANGED
@@ -78,6 +78,18 @@ _wandb:
|
|
78 |
5: 1
|
79 |
6:
|
80 |
- 1
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
81 |
vocab_size:
|
82 |
desc: null
|
83 |
value: 30522
|
|
|
78 |
5: 1
|
79 |
6:
|
80 |
- 1
|
81 |
+
- 1: train/loss
|
82 |
+
5: 1
|
83 |
+
6:
|
84 |
+
- 1
|
85 |
+
- 1: train/grad_norm
|
86 |
+
5: 1
|
87 |
+
6:
|
88 |
+
- 1
|
89 |
+
- 1: train/learning_rate
|
90 |
+
5: 1
|
91 |
+
6:
|
92 |
+
- 1
|
93 |
vocab_size:
|
94 |
desc: null
|
95 |
value: 30522
|
wandb/run-20240924_172630-x9iddikd/files/output.log
CHANGED
@@ -19,3 +19,5 @@ dados_tokenizados:
|
|
19 |
warnings.warn(
|
20 |
Some weights of DistilBertForSequenceClassification were not initialized from the model checkpoint at distilbert/distilbert-base-uncased and are newly initialized: ['classifier.bias', 'classifier.weight', 'pre_classifier.bias', 'pre_classifier.weight']
|
21 |
You should probably TRAIN this model on a down-stream task to be able to use it for predictions and inference.
|
|
|
|
|
|
19 |
warnings.warn(
|
20 |
Some weights of DistilBertForSequenceClassification were not initialized from the model checkpoint at distilbert/distilbert-base-uncased and are newly initialized: ['classifier.bias', 'classifier.weight', 'pre_classifier.bias', 'pre_classifier.weight']
|
21 |
You should probably TRAIN this model on a down-stream task to be able to use it for predictions and inference.
|
22 |
+
{'eval_loss': 0.21834564208984375, 'eval_accuracy': 0.938, 'eval_runtime': 30.3683, 'eval_samples_per_second': 32.929, 'eval_steps_per_second': 2.075, 'epoch': 1.0}
|
23 |
+
{'loss': 0.2031, 'grad_norm': 1.1480563879013062, 'learning_rate': 1.2e-05, 'epoch': 2.0}
|
wandb/run-20240924_172630-x9iddikd/files/wandb-summary.json
CHANGED
@@ -1 +1 @@
|
|
1 |
-
{"eval/loss": 0.
|
|
|
1 |
+
{"eval/loss": 0.19427122175693512, "eval/accuracy": 0.938, "eval/runtime": 42.2287, "eval/samples_per_second": 23.681, "eval/steps_per_second": 1.492, "train/epoch": 2.0, "train/global_step": 500, "_timestamp": 1727212675.9347782, "_runtime": 3085.0618691444397, "_step": 2, "train/loss": 0.2031, "train/grad_norm": 1.1480563879013062, "train/learning_rate": 1.2e-05}
|
wandb/run-20240924_172630-x9iddikd/logs/debug-internal.log
CHANGED
The diff for this file is too large to render.
See raw diff
|
|
wandb/run-20240924_172630-x9iddikd/run-x9iddikd.wandb
CHANGED
Binary files a/wandb/run-20240924_172630-x9iddikd/run-x9iddikd.wandb and b/wandb/run-20240924_172630-x9iddikd/run-x9iddikd.wandb differ
|
|