0xsuid commited on
Commit
0f59e3f
·
1 Parent(s): 2f85276

Upload experiments/ with huggingface_hub

Browse files
Files changed (19) hide show
  1. experiments/.gitkeep +0 -0
  2. experiments/2023-02-22-5f8754af240241f4c78ba924918b0bfb26f26a17b0e1799c8a046eb3f6dd6e94/configs.json +26 -0
  3. experiments/2023-02-22-5f8754af240241f4c78ba924918b0bfb26f26a17b0e1799c8a046eb3f6dd6e94/deepspeed.json +48 -0
  4. experiments/2023-02-22-5f8754af240241f4c78ba924918b0bfb26f26a17b0e1799c8a046eb3f6dd6e94/final_checkpoint/config.json +54 -0
  5. experiments/2023-02-22-5f8754af240241f4c78ba924918b0bfb26f26a17b0e1799c8a046eb3f6dd6e94/final_checkpoint/pytorch_model.bin +3 -0
  6. experiments/2023-02-22-5f8754af240241f4c78ba924918b0bfb26f26a17b0e1799c8a046eb3f6dd6e94/final_checkpoint/tokenizer/merges.txt +0 -0
  7. experiments/2023-02-22-5f8754af240241f4c78ba924918b0bfb26f26a17b0e1799c8a046eb3f6dd6e94/final_checkpoint/tokenizer/special_tokens_map.json +24 -0
  8. experiments/2023-02-22-5f8754af240241f4c78ba924918b0bfb26f26a17b0e1799c8a046eb3f6dd6e94/final_checkpoint/tokenizer/tokenizer.json +0 -0
  9. experiments/2023-02-22-5f8754af240241f4c78ba924918b0bfb26f26a17b0e1799c8a046eb3f6dd6e94/final_checkpoint/tokenizer/tokenizer_config.json +34 -0
  10. experiments/2023-02-22-5f8754af240241f4c78ba924918b0bfb26f26a17b0e1799c8a046eb3f6dd6e94/final_checkpoint/tokenizer/vocab.json +0 -0
  11. experiments/2023-02-22-5f8754af240241f4c78ba924918b0bfb26f26a17b0e1799c8a046eb3f6dd6e94/output.log +0 -0
  12. experiments/2023-02-22-5f8754af240241f4c78ba924918b0bfb26f26a17b0e1799c8a046eb3f6dd6e94/trainer_final_checkpoint/config.json +54 -0
  13. experiments/2023-02-22-5f8754af240241f4c78ba924918b0bfb26f26a17b0e1799c8a046eb3f6dd6e94/trainer_final_checkpoint/merges.txt +0 -0
  14. experiments/2023-02-22-5f8754af240241f4c78ba924918b0bfb26f26a17b0e1799c8a046eb3f6dd6e94/trainer_final_checkpoint/pytorch_model.bin +3 -0
  15. experiments/2023-02-22-5f8754af240241f4c78ba924918b0bfb26f26a17b0e1799c8a046eb3f6dd6e94/trainer_final_checkpoint/special_tokens_map.json +24 -0
  16. experiments/2023-02-22-5f8754af240241f4c78ba924918b0bfb26f26a17b0e1799c8a046eb3f6dd6e94/trainer_final_checkpoint/tokenizer.json +0 -0
  17. experiments/2023-02-22-5f8754af240241f4c78ba924918b0bfb26f26a17b0e1799c8a046eb3f6dd6e94/trainer_final_checkpoint/tokenizer_config.json +34 -0
  18. experiments/2023-02-22-5f8754af240241f4c78ba924918b0bfb26f26a17b0e1799c8a046eb3f6dd6e94/trainer_final_checkpoint/training_args.bin +3 -0
  19. experiments/2023-02-22-5f8754af240241f4c78ba924918b0bfb26f26a17b0e1799c8a046eb3f6dd6e94/trainer_final_checkpoint/vocab.json +0 -0
experiments/.gitkeep ADDED
File without changes
experiments/2023-02-22-5f8754af240241f4c78ba924918b0bfb26f26a17b0e1799c8a046eb3f6dd6e94/configs.json ADDED
@@ -0,0 +1,26 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "output_dir": "./results",
3
+ "evaluation_strategy": "no",
4
+ "do_eval": false,
5
+ "eval_steps": 0,
6
+ "log_level": "info",
7
+ "logging_first_step": true,
8
+ "logging_steps": 5,
9
+ "logging_dir": "./logs",
10
+ "save_steps": 150,
11
+ "save_total_limit": 1,
12
+ "num_train_epochs": 10,
13
+ "per_device_train_batch_size": 6,
14
+ "optim": "adamw_torch",
15
+ "gradient_accumulation_steps": 4,
16
+ "dataloader_drop_last": true,
17
+ "warmup_steps": 1000,
18
+ "weight_decay": 0.1,
19
+ "learning_rate": 0.0001,
20
+ "deepspeed": "deepspeed.json",
21
+ "local_rank": 0,
22
+ "total_gpus": 8,
23
+ "v_cpus": 64,
24
+ "total_memory_in_gb": 257604.9765625,
25
+ "dataset_limit": 0
26
+ }
experiments/2023-02-22-5f8754af240241f4c78ba924918b0bfb26f26a17b0e1799c8a046eb3f6dd6e94/deepspeed.json ADDED
@@ -0,0 +1,48 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "optimizer": {
3
+ "type": "AdamW",
4
+ "params": {
5
+ "lr": 1e-4,
6
+ "betas": [ 0.9, 0.999 ],
7
+ "eps": 1e-8,
8
+ "weight_decay": 0.1
9
+ }
10
+ },
11
+
12
+ "scheduler": {
13
+ "type": "WarmupLR",
14
+ "params": {
15
+ "warmup_min_lr": 1e-9,
16
+ "warmup_max_lr": 1e-4,
17
+ "warmup_num_steps": 1000
18
+ }
19
+ },
20
+
21
+ "zero_optimization": {
22
+ "stage": 2,
23
+ "offload_optimizer": {
24
+ "device": "cpu",
25
+ "pin_memory": true
26
+ },
27
+ "allgather_partitions": true,
28
+ "allgather_bucket_size": 5e8,
29
+ "overlap_comm": true,
30
+ "reduce_scatter": true,
31
+ "reduce_bucket_size": 5e8,
32
+ "contiguous_gradients": true
33
+ },
34
+
35
+ "tensorboard": {
36
+ "enabled": true,
37
+ "output_path": "logs/",
38
+ "job_name": "train_neo"
39
+ },
40
+
41
+ "zero_allow_untested_optimizer": true,
42
+ "gradient_accumulation_steps": "auto",
43
+ "gradient_clipping": "auto",
44
+ "steps_per_print": 2000,
45
+ "train_batch_size": "auto",
46
+ "train_micro_batch_size_per_gpu": "auto",
47
+ "wall_clock_breakdown": false
48
+ }
experiments/2023-02-22-5f8754af240241f4c78ba924918b0bfb26f26a17b0e1799c8a046eb3f6dd6e94/final_checkpoint/config.json ADDED
@@ -0,0 +1,54 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "_name_or_path": "EleutherAI/gpt-neo-125M",
3
+ "activation_function": "gelu_new",
4
+ "architectures": [
5
+ "GPTNeoForCausalLM"
6
+ ],
7
+ "attention_dropout": 0,
8
+ "attention_layers": [
9
+ "global",
10
+ "local",
11
+ "global",
12
+ "local",
13
+ "global",
14
+ "local",
15
+ "global",
16
+ "local",
17
+ "global",
18
+ "local",
19
+ "global",
20
+ "local"
21
+ ],
22
+ "attention_types": [
23
+ [
24
+ [
25
+ "global",
26
+ "local"
27
+ ],
28
+ 6
29
+ ]
30
+ ],
31
+ "bos_token_id": 50256,
32
+ "embed_dropout": 0,
33
+ "eos_token_id": 50256,
34
+ "gradient_checkpointing": false,
35
+ "hidden_size": 768,
36
+ "initializer_range": 0.02,
37
+ "intermediate_size": null,
38
+ "layer_norm_epsilon": 1e-05,
39
+ "max_position_embeddings": 2048,
40
+ "model_type": "gpt_neo",
41
+ "num_heads": 12,
42
+ "num_layers": 12,
43
+ "resid_dropout": 0,
44
+ "summary_activation": null,
45
+ "summary_first_dropout": 0.1,
46
+ "summary_proj_to_labels": true,
47
+ "summary_type": "cls_index",
48
+ "summary_use_proj": true,
49
+ "torch_dtype": "float32",
50
+ "transformers_version": "4.24.0",
51
+ "use_cache": true,
52
+ "vocab_size": 50257,
53
+ "window_size": 256
54
+ }
experiments/2023-02-22-5f8754af240241f4c78ba924918b0bfb26f26a17b0e1799c8a046eb3f6dd6e94/final_checkpoint/pytorch_model.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:6589651dd8c9588a8faa97daec7814ffcfe998ba2e481eabdcb3e75405761d5d
3
+ size 551154684
experiments/2023-02-22-5f8754af240241f4c78ba924918b0bfb26f26a17b0e1799c8a046eb3f6dd6e94/final_checkpoint/tokenizer/merges.txt ADDED
The diff for this file is too large to render. See raw diff
 
experiments/2023-02-22-5f8754af240241f4c78ba924918b0bfb26f26a17b0e1799c8a046eb3f6dd6e94/final_checkpoint/tokenizer/special_tokens_map.json ADDED
@@ -0,0 +1,24 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "bos_token": {
3
+ "content": "<|endoftext|>",
4
+ "lstrip": false,
5
+ "normalized": true,
6
+ "rstrip": false,
7
+ "single_word": false
8
+ },
9
+ "eos_token": {
10
+ "content": "<|endoftext|>",
11
+ "lstrip": false,
12
+ "normalized": true,
13
+ "rstrip": false,
14
+ "single_word": false
15
+ },
16
+ "pad_token": "<|endoftext|>",
17
+ "unk_token": {
18
+ "content": "<|endoftext|>",
19
+ "lstrip": false,
20
+ "normalized": true,
21
+ "rstrip": false,
22
+ "single_word": false
23
+ }
24
+ }
experiments/2023-02-22-5f8754af240241f4c78ba924918b0bfb26f26a17b0e1799c8a046eb3f6dd6e94/final_checkpoint/tokenizer/tokenizer.json ADDED
The diff for this file is too large to render. See raw diff
 
experiments/2023-02-22-5f8754af240241f4c78ba924918b0bfb26f26a17b0e1799c8a046eb3f6dd6e94/final_checkpoint/tokenizer/tokenizer_config.json ADDED
@@ -0,0 +1,34 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "add_bos_token": false,
3
+ "add_prefix_space": false,
4
+ "bos_token": {
5
+ "__type": "AddedToken",
6
+ "content": "<|endoftext|>",
7
+ "lstrip": false,
8
+ "normalized": true,
9
+ "rstrip": false,
10
+ "single_word": false
11
+ },
12
+ "eos_token": {
13
+ "__type": "AddedToken",
14
+ "content": "<|endoftext|>",
15
+ "lstrip": false,
16
+ "normalized": true,
17
+ "rstrip": false,
18
+ "single_word": false
19
+ },
20
+ "errors": "replace",
21
+ "model_max_length": 2048,
22
+ "name_or_path": "EleutherAI/gpt-neo-125M",
23
+ "pad_token": null,
24
+ "special_tokens_map_file": null,
25
+ "tokenizer_class": "GPT2Tokenizer",
26
+ "unk_token": {
27
+ "__type": "AddedToken",
28
+ "content": "<|endoftext|>",
29
+ "lstrip": false,
30
+ "normalized": true,
31
+ "rstrip": false,
32
+ "single_word": false
33
+ }
34
+ }
experiments/2023-02-22-5f8754af240241f4c78ba924918b0bfb26f26a17b0e1799c8a046eb3f6dd6e94/final_checkpoint/tokenizer/vocab.json ADDED
The diff for this file is too large to render. See raw diff
 
experiments/2023-02-22-5f8754af240241f4c78ba924918b0bfb26f26a17b0e1799c8a046eb3f6dd6e94/output.log ADDED
The diff for this file is too large to render. See raw diff
 
experiments/2023-02-22-5f8754af240241f4c78ba924918b0bfb26f26a17b0e1799c8a046eb3f6dd6e94/trainer_final_checkpoint/config.json ADDED
@@ -0,0 +1,54 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "_name_or_path": "EleutherAI/gpt-neo-125M",
3
+ "activation_function": "gelu_new",
4
+ "architectures": [
5
+ "GPTNeoForCausalLM"
6
+ ],
7
+ "attention_dropout": 0,
8
+ "attention_layers": [
9
+ "global",
10
+ "local",
11
+ "global",
12
+ "local",
13
+ "global",
14
+ "local",
15
+ "global",
16
+ "local",
17
+ "global",
18
+ "local",
19
+ "global",
20
+ "local"
21
+ ],
22
+ "attention_types": [
23
+ [
24
+ [
25
+ "global",
26
+ "local"
27
+ ],
28
+ 6
29
+ ]
30
+ ],
31
+ "bos_token_id": 50256,
32
+ "embed_dropout": 0,
33
+ "eos_token_id": 50256,
34
+ "gradient_checkpointing": false,
35
+ "hidden_size": 768,
36
+ "initializer_range": 0.02,
37
+ "intermediate_size": null,
38
+ "layer_norm_epsilon": 1e-05,
39
+ "max_position_embeddings": 2048,
40
+ "model_type": "gpt_neo",
41
+ "num_heads": 12,
42
+ "num_layers": 12,
43
+ "resid_dropout": 0,
44
+ "summary_activation": null,
45
+ "summary_first_dropout": 0.1,
46
+ "summary_proj_to_labels": true,
47
+ "summary_type": "cls_index",
48
+ "summary_use_proj": true,
49
+ "torch_dtype": "float32",
50
+ "transformers_version": "4.24.0",
51
+ "use_cache": true,
52
+ "vocab_size": 50257,
53
+ "window_size": 256
54
+ }
experiments/2023-02-22-5f8754af240241f4c78ba924918b0bfb26f26a17b0e1799c8a046eb3f6dd6e94/trainer_final_checkpoint/merges.txt ADDED
The diff for this file is too large to render. See raw diff
 
experiments/2023-02-22-5f8754af240241f4c78ba924918b0bfb26f26a17b0e1799c8a046eb3f6dd6e94/trainer_final_checkpoint/pytorch_model.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:6589651dd8c9588a8faa97daec7814ffcfe998ba2e481eabdcb3e75405761d5d
3
+ size 551154684
experiments/2023-02-22-5f8754af240241f4c78ba924918b0bfb26f26a17b0e1799c8a046eb3f6dd6e94/trainer_final_checkpoint/special_tokens_map.json ADDED
@@ -0,0 +1,24 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "bos_token": {
3
+ "content": "<|endoftext|>",
4
+ "lstrip": false,
5
+ "normalized": true,
6
+ "rstrip": false,
7
+ "single_word": false
8
+ },
9
+ "eos_token": {
10
+ "content": "<|endoftext|>",
11
+ "lstrip": false,
12
+ "normalized": true,
13
+ "rstrip": false,
14
+ "single_word": false
15
+ },
16
+ "pad_token": "<|endoftext|>",
17
+ "unk_token": {
18
+ "content": "<|endoftext|>",
19
+ "lstrip": false,
20
+ "normalized": true,
21
+ "rstrip": false,
22
+ "single_word": false
23
+ }
24
+ }
experiments/2023-02-22-5f8754af240241f4c78ba924918b0bfb26f26a17b0e1799c8a046eb3f6dd6e94/trainer_final_checkpoint/tokenizer.json ADDED
The diff for this file is too large to render. See raw diff
 
experiments/2023-02-22-5f8754af240241f4c78ba924918b0bfb26f26a17b0e1799c8a046eb3f6dd6e94/trainer_final_checkpoint/tokenizer_config.json ADDED
@@ -0,0 +1,34 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "add_bos_token": false,
3
+ "add_prefix_space": false,
4
+ "bos_token": {
5
+ "__type": "AddedToken",
6
+ "content": "<|endoftext|>",
7
+ "lstrip": false,
8
+ "normalized": true,
9
+ "rstrip": false,
10
+ "single_word": false
11
+ },
12
+ "eos_token": {
13
+ "__type": "AddedToken",
14
+ "content": "<|endoftext|>",
15
+ "lstrip": false,
16
+ "normalized": true,
17
+ "rstrip": false,
18
+ "single_word": false
19
+ },
20
+ "errors": "replace",
21
+ "model_max_length": 2048,
22
+ "name_or_path": "EleutherAI/gpt-neo-125M",
23
+ "pad_token": null,
24
+ "special_tokens_map_file": null,
25
+ "tokenizer_class": "GPT2Tokenizer",
26
+ "unk_token": {
27
+ "__type": "AddedToken",
28
+ "content": "<|endoftext|>",
29
+ "lstrip": false,
30
+ "normalized": true,
31
+ "rstrip": false,
32
+ "single_word": false
33
+ }
34
+ }
experiments/2023-02-22-5f8754af240241f4c78ba924918b0bfb26f26a17b0e1799c8a046eb3f6dd6e94/trainer_final_checkpoint/training_args.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:69255842ed78e689b179c20bae1499576b967ef499d84d9ec162b162d779e910
3
+ size 4475
experiments/2023-02-22-5f8754af240241f4c78ba924918b0bfb26f26a17b0e1799c8a046eb3f6dd6e94/trainer_final_checkpoint/vocab.json ADDED
The diff for this file is too large to render. See raw diff