yhavinga commited on
Commit
68faee3
1 Parent(s): 9fd0f29

Saving at step 2840k, loss 1.108, acc 0.747

Browse files
Files changed (6) hide show
  1. eval_results.json +4 -0
  2. flax_model.msgpack +1 -1
  3. info.txt +9 -0
  4. opt_state.msgpack +1 -1
  5. run.sh +38 -0
  6. training_state.json +1 -1
eval_results.json ADDED
@@ -0,0 +1,4 @@
 
 
 
 
 
1
+ {
2
+ "eval_accuracy": 0.7469202280044556,
3
+ "eval_loss": 1.1080663204193115
4
+ }
flax_model.msgpack CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:be580d25b36fead261bee39bb04610e55d681c9880ea51883d2747d24f950c28
3
  size 990170015
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:b38f40c38ee1b037c8c509dd349a3c8d3c16870fa0571f9d947c7cd1680ebd9d
3
  size 990170015
info.txt ADDED
@@ -0,0 +1,9 @@
 
 
 
 
 
 
 
 
 
 
1
+ INFO:__main__: Optimizer = adafactor
2
+ INFO:__main__: Learning rate (peak) = 0.005
3
+ INFO:__main__: Num examples = 31519126
4
+ INFO:__main__: Num tokenized group examples 36347268
5
+ INFO:__main__: Num Epochs = 10
6
+ INFO:__main__: Instantaneous batch size per device = 16
7
+ INFO:__main__: Total train batch size (w. parallel & grad accum) = 128
8
+ INFO:__main__: Steps per epoch = 283963
9
+ INFO:__main__: Total optimization steps = 2839630
opt_state.msgpack CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:81e11965da1594b414d17cf95b0dc6e0e62f041352c61969a6a8a8141351df5e
3
  size 2184331
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:5871414180a92600a235367a2d7dd7c6a9ca06ac3b2311a6d87203383992474c
3
  size 2184331
run.sh ADDED
@@ -0,0 +1,38 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ export HF_PROJECT="t5-v1_1-base-dutch-english-cased"
2
+ export DATASET="yhavinga/mc4_nl_cleaned" # Name of the dataset in the Huggingface Hub
3
+ export DATASET_CONFIG="small_en_nl" # Config of the dataset in the Huggingface Hub
4
+ export DATASET_SPLIT="train" # Split to use for training tokenizer and model
5
+ export CONFIG_NAME="yhavinga/t5-v1.1-base-dutch-cased"
6
+ export TOKENIZER_NAME="yhavinga/net5-v1.1-base-cased-500"
7
+ export MODEL_PATH="${HOME}/data/${HF_PROJECT}" # Path to the model
8
+
9
+
10
+ python3 ../train/run_t5_mlm_flax_pmap.py \
11
+ --output_dir="${MODEL_PATH}" \
12
+ --model_type="t5" \
13
+ --config_name="${CONFIG_NAME}" \
14
+ --tokenizer_name="${TOKENIZER_NAME}" \
15
+ --auth_token="$(cat ~/.huggingface/token)" \
16
+ --preprocessing_num_workers="96" \
17
+ --do_train --do_eval \
18
+ --dataset_name="${DATASET}" \
19
+ --dataset_config_name="${DATASET_CONFIG}" \
20
+ --max_seq_length="512" \
21
+ --per_device_train_batch_size="16" \
22
+ --per_device_eval_batch_size="16" \
23
+ --optim="adafactor" \
24
+ --learning_rate="0.005" \
25
+ --lr_decay="exponential" \
26
+ --lr_transition_steps="300000" \
27
+ --lr_decay_rate="0.7" \
28
+ --lr_staircase="false" \
29
+ --overwrite_output_dir \
30
+ --num_train_epochs="10" \
31
+ --logging_steps="200" \
32
+ --save_steps="10000" \
33
+ --eval_steps="1250" \
34
+ --warmup_steps="10000" \
35
+ --validation_split_count="15000" \
36
+ --wandb_project="t5-v1_1-dutch-english" \
37
+ --wandb_job_type="pmap" \
38
+ --resume_from_checkpoint="${MODEL_PATH}"
training_state.json CHANGED
@@ -1 +1 @@
1
- {"step": 2119986}
 
1
+ {"step": 2829975}