lehduong commited on
Commit
04e54ee
1 Parent(s): 81c3d13

Upload folder using huggingface_hub

Browse files
README.md CHANGED
@@ -1,3 +1,2 @@
1
- ---
2
- license: cc-by-nc-4.0
3
- ---
 
1
+ # OneDiffusion
2
+
 
scheduler/scheduler_config.json ADDED
@@ -0,0 +1,12 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "_class_name": "FlowMatchEulerDiscreteScheduler",
3
+ "_diffusers_version": "0.29.0.dev0",
4
+ "base_image_seq_len": 256,
5
+ "base_shift": 0.5,
6
+ "max_image_seq_len": 4096,
7
+ "max_shift": 1.15,
8
+ "num_train_timesteps": 1000,
9
+ "shift": 3.0,
10
+ "use_dynamic_shifting": true
11
+ }
12
+
text_encoder/config.json ADDED
@@ -0,0 +1,25 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "_name_or_path": "/home/patrick/t5/t5-v1_1-xl",
3
+ "architectures": [
4
+ "T5ForConditionalGeneration"
5
+ ],
6
+ "d_ff": 5120,
7
+ "d_kv": 64,
8
+ "d_model": 2048,
9
+ "decoder_start_token_id": 0,
10
+ "dropout_rate": 0.1,
11
+ "eos_token_id": 1,
12
+ "feed_forward_proj": "gated-gelu",
13
+ "initializer_factor": 1.0,
14
+ "is_encoder_decoder": true,
15
+ "layer_norm_epsilon": 1e-06,
16
+ "model_type": "t5",
17
+ "num_decoder_layers": 24,
18
+ "num_heads": 32,
19
+ "num_layers": 24,
20
+ "output_past": true,
21
+ "pad_token_id": 0,
22
+ "relative_attention_num_buckets": 32,
23
+ "tie_word_embeddings": false,
24
+ "vocab_size": 32128
25
+ }
text_encoder/pytorch_model.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:72be4ab4cb05720c91e72995ad26fb5e2785fc1648c80ad0561c45ce432447fc
3
+ size 11399285345
tokenizer/special_tokens_map.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"eos_token": "</s>", "unk_token": "<unk>", "pad_token": "<pad>", "additional_special_tokens": ["<extra_id_0>", "<extra_id_1>", "<extra_id_2>", "<extra_id_3>", "<extra_id_4>", "<extra_id_5>", "<extra_id_6>", "<extra_id_7>", "<extra_id_8>", "<extra_id_9>", "<extra_id_10>", "<extra_id_11>", "<extra_id_12>", "<extra_id_13>", "<extra_id_14>", "<extra_id_15>", "<extra_id_16>", "<extra_id_17>", "<extra_id_18>", "<extra_id_19>", "<extra_id_20>", "<extra_id_21>", "<extra_id_22>", "<extra_id_23>", "<extra_id_24>", "<extra_id_25>", "<extra_id_26>", "<extra_id_27>", "<extra_id_28>", "<extra_id_29>", "<extra_id_30>", "<extra_id_31>", "<extra_id_32>", "<extra_id_33>", "<extra_id_34>", "<extra_id_35>", "<extra_id_36>", "<extra_id_37>", "<extra_id_38>", "<extra_id_39>", "<extra_id_40>", "<extra_id_41>", "<extra_id_42>", "<extra_id_43>", "<extra_id_44>", "<extra_id_45>", "<extra_id_46>", "<extra_id_47>", "<extra_id_48>", "<extra_id_49>", "<extra_id_50>", "<extra_id_51>", "<extra_id_52>", "<extra_id_53>", "<extra_id_54>", "<extra_id_55>", "<extra_id_56>", "<extra_id_57>", "<extra_id_58>", "<extra_id_59>", "<extra_id_60>", "<extra_id_61>", "<extra_id_62>", "<extra_id_63>", "<extra_id_64>", "<extra_id_65>", "<extra_id_66>", "<extra_id_67>", "<extra_id_68>", "<extra_id_69>", "<extra_id_70>", "<extra_id_71>", "<extra_id_72>", "<extra_id_73>", "<extra_id_74>", "<extra_id_75>", "<extra_id_76>", "<extra_id_77>", "<extra_id_78>", "<extra_id_79>", "<extra_id_80>", "<extra_id_81>", "<extra_id_82>", "<extra_id_83>", "<extra_id_84>", "<extra_id_85>", "<extra_id_86>", "<extra_id_87>", "<extra_id_88>", "<extra_id_89>", "<extra_id_90>", "<extra_id_91>", "<extra_id_92>", "<extra_id_93>", "<extra_id_94>", "<extra_id_95>", "<extra_id_96>", "<extra_id_97>", "<extra_id_98>", "<extra_id_99>"]}
tokenizer/spiece.model ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:d60acb128cf7b7f2536e8f38a5b18a05535c9e14c7a355904270e15b0945ea86
3
+ size 791656
tokenizer/tokenizer_config.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"eos_token": "</s>", "unk_token": "<unk>", "pad_token": "<pad>", "extra_ids": 100, "additional_special_tokens": ["<extra_id_0>", "<extra_id_1>", "<extra_id_2>", "<extra_id_3>", "<extra_id_4>", "<extra_id_5>", "<extra_id_6>", "<extra_id_7>", "<extra_id_8>", "<extra_id_9>", "<extra_id_10>", "<extra_id_11>", "<extra_id_12>", "<extra_id_13>", "<extra_id_14>", "<extra_id_15>", "<extra_id_16>", "<extra_id_17>", "<extra_id_18>", "<extra_id_19>", "<extra_id_20>", "<extra_id_21>", "<extra_id_22>", "<extra_id_23>", "<extra_id_24>", "<extra_id_25>", "<extra_id_26>", "<extra_id_27>", "<extra_id_28>", "<extra_id_29>", "<extra_id_30>", "<extra_id_31>", "<extra_id_32>", "<extra_id_33>", "<extra_id_34>", "<extra_id_35>", "<extra_id_36>", "<extra_id_37>", "<extra_id_38>", "<extra_id_39>", "<extra_id_40>", "<extra_id_41>", "<extra_id_42>", "<extra_id_43>", "<extra_id_44>", "<extra_id_45>", "<extra_id_46>", "<extra_id_47>", "<extra_id_48>", "<extra_id_49>", "<extra_id_50>", "<extra_id_51>", "<extra_id_52>", "<extra_id_53>", "<extra_id_54>", "<extra_id_55>", "<extra_id_56>", "<extra_id_57>", "<extra_id_58>", "<extra_id_59>", "<extra_id_60>", "<extra_id_61>", "<extra_id_62>", "<extra_id_63>", "<extra_id_64>", "<extra_id_65>", "<extra_id_66>", "<extra_id_67>", "<extra_id_68>", "<extra_id_69>", "<extra_id_70>", "<extra_id_71>", "<extra_id_72>", "<extra_id_73>", "<extra_id_74>", "<extra_id_75>", "<extra_id_76>", "<extra_id_77>", "<extra_id_78>", "<extra_id_79>", "<extra_id_80>", "<extra_id_81>", "<extra_id_82>", "<extra_id_83>", "<extra_id_84>", "<extra_id_85>", "<extra_id_86>", "<extra_id_87>", "<extra_id_88>", "<extra_id_89>", "<extra_id_90>", "<extra_id_91>", "<extra_id_92>", "<extra_id_93>", "<extra_id_94>", "<extra_id_95>", "<extra_id_96>", "<extra_id_97>", "<extra_id_98>", "<extra_id_99>"], "model_max_length": 512, "name_or_path": "t5-small"}
transformer/config.json ADDED
@@ -0,0 +1,41 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "_class_name": "NextDiT",
3
+ "_diffusers_version": "0.30.3",
4
+ "caption_channels": 2048,
5
+ "decay": 0.9999,
6
+ "depth": 16,
7
+ "ffn_dim_multiplier": null,
8
+ "forward_dtype": "bfloat16",
9
+ "hidden_size": 3840,
10
+ "in_channels": 16,
11
+ "input_size": [
12
+ 1,
13
+ 128,
14
+ 128
15
+ ],
16
+ "inv_gamma": 1.0,
17
+ "min_decay": 0.0,
18
+ "model_max_length": 512,
19
+ "multiple_of": 256,
20
+ "norm_eps": 1e-05,
21
+ "norm_type": "rms",
22
+ "num_heads": 32,
23
+ "num_kv_heads": 8,
24
+ "optimization_step": 4000,
25
+ "param_dtype": "float32",
26
+ "patch_size": [
27
+ 1,
28
+ 2,
29
+ 2
30
+ ],
31
+ "power": 0.6666666666666666,
32
+ "pred_sigma": false,
33
+ "qk_norm": true,
34
+ "remat_attention": "",
35
+ "remat_block": "",
36
+ "remat_mlp": "",
37
+ "rotary_max_length": 128,
38
+ "rotary_max_length_t": null,
39
+ "update_after_step": 0,
40
+ "use_ema_warmup": false
41
+ }
transformer/diffusion_pytorch_model-00001-of-00002.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:17f844195518e70960b2bc50af3f7aae51ff2e0ba59ce5b6e6b0d922eb74ce29
3
+ size 9951824992
transformer/diffusion_pytorch_model-00002-of-00002.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:5c8953dc88e536b1bd8de4cdede32fac340a5145fe9b0c7cd6e2400431880333
3
+ size 1249771672
transformer/diffusion_pytorch_model.safetensors.index.json ADDED
@@ -0,0 +1,389 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "metadata": {
3
+ "total_size": 11201556736
4
+ },
5
+ "weight_map": {
6
+ "final_layer.adaLN_modulation.1.bias": "diffusion_pytorch_model-00002-of-00002.safetensors",
7
+ "final_layer.adaLN_modulation.1.weight": "diffusion_pytorch_model-00002-of-00002.safetensors",
8
+ "final_layer.linear.bias": "diffusion_pytorch_model-00002-of-00002.safetensors",
9
+ "final_layer.linear.weight": "diffusion_pytorch_model-00002-of-00002.safetensors",
10
+ "layers.0.adaLN_modulation.1.bias": "diffusion_pytorch_model-00001-of-00002.safetensors",
11
+ "layers.0.adaLN_modulation.1.weight": "diffusion_pytorch_model-00001-of-00002.safetensors",
12
+ "layers.0.attention.gate": "diffusion_pytorch_model-00001-of-00002.safetensors",
13
+ "layers.0.attention.k_norm.bias": "diffusion_pytorch_model-00001-of-00002.safetensors",
14
+ "layers.0.attention.k_norm.weight": "diffusion_pytorch_model-00001-of-00002.safetensors",
15
+ "layers.0.attention.ky_norm.bias": "diffusion_pytorch_model-00001-of-00002.safetensors",
16
+ "layers.0.attention.ky_norm.weight": "diffusion_pytorch_model-00001-of-00002.safetensors",
17
+ "layers.0.attention.q_norm.bias": "diffusion_pytorch_model-00001-of-00002.safetensors",
18
+ "layers.0.attention.q_norm.weight": "diffusion_pytorch_model-00001-of-00002.safetensors",
19
+ "layers.0.attention.wk.weight": "diffusion_pytorch_model-00001-of-00002.safetensors",
20
+ "layers.0.attention.wk_y.weight": "diffusion_pytorch_model-00001-of-00002.safetensors",
21
+ "layers.0.attention.wo.weight": "diffusion_pytorch_model-00001-of-00002.safetensors",
22
+ "layers.0.attention.wq.weight": "diffusion_pytorch_model-00001-of-00002.safetensors",
23
+ "layers.0.attention.wv.weight": "diffusion_pytorch_model-00001-of-00002.safetensors",
24
+ "layers.0.attention.wv_y.weight": "diffusion_pytorch_model-00001-of-00002.safetensors",
25
+ "layers.0.attention_norm1.weight": "diffusion_pytorch_model-00001-of-00002.safetensors",
26
+ "layers.0.attention_norm2.weight": "diffusion_pytorch_model-00001-of-00002.safetensors",
27
+ "layers.0.attention_y_norm.weight": "diffusion_pytorch_model-00001-of-00002.safetensors",
28
+ "layers.0.feed_forward.w1.weight": "diffusion_pytorch_model-00001-of-00002.safetensors",
29
+ "layers.0.feed_forward.w2.weight": "diffusion_pytorch_model-00001-of-00002.safetensors",
30
+ "layers.0.feed_forward.w3.weight": "diffusion_pytorch_model-00001-of-00002.safetensors",
31
+ "layers.0.ffn_norm1.weight": "diffusion_pytorch_model-00001-of-00002.safetensors",
32
+ "layers.0.ffn_norm2.weight": "diffusion_pytorch_model-00001-of-00002.safetensors",
33
+ "layers.1.adaLN_modulation.1.bias": "diffusion_pytorch_model-00001-of-00002.safetensors",
34
+ "layers.1.adaLN_modulation.1.weight": "diffusion_pytorch_model-00001-of-00002.safetensors",
35
+ "layers.1.attention.gate": "diffusion_pytorch_model-00001-of-00002.safetensors",
36
+ "layers.1.attention.k_norm.bias": "diffusion_pytorch_model-00001-of-00002.safetensors",
37
+ "layers.1.attention.k_norm.weight": "diffusion_pytorch_model-00001-of-00002.safetensors",
38
+ "layers.1.attention.ky_norm.bias": "diffusion_pytorch_model-00001-of-00002.safetensors",
39
+ "layers.1.attention.ky_norm.weight": "diffusion_pytorch_model-00001-of-00002.safetensors",
40
+ "layers.1.attention.q_norm.bias": "diffusion_pytorch_model-00001-of-00002.safetensors",
41
+ "layers.1.attention.q_norm.weight": "diffusion_pytorch_model-00001-of-00002.safetensors",
42
+ "layers.1.attention.wk.weight": "diffusion_pytorch_model-00001-of-00002.safetensors",
43
+ "layers.1.attention.wk_y.weight": "diffusion_pytorch_model-00001-of-00002.safetensors",
44
+ "layers.1.attention.wo.weight": "diffusion_pytorch_model-00001-of-00002.safetensors",
45
+ "layers.1.attention.wq.weight": "diffusion_pytorch_model-00001-of-00002.safetensors",
46
+ "layers.1.attention.wv.weight": "diffusion_pytorch_model-00001-of-00002.safetensors",
47
+ "layers.1.attention.wv_y.weight": "diffusion_pytorch_model-00001-of-00002.safetensors",
48
+ "layers.1.attention_norm1.weight": "diffusion_pytorch_model-00001-of-00002.safetensors",
49
+ "layers.1.attention_norm2.weight": "diffusion_pytorch_model-00001-of-00002.safetensors",
50
+ "layers.1.attention_y_norm.weight": "diffusion_pytorch_model-00001-of-00002.safetensors",
51
+ "layers.1.feed_forward.w1.weight": "diffusion_pytorch_model-00001-of-00002.safetensors",
52
+ "layers.1.feed_forward.w2.weight": "diffusion_pytorch_model-00001-of-00002.safetensors",
53
+ "layers.1.feed_forward.w3.weight": "diffusion_pytorch_model-00001-of-00002.safetensors",
54
+ "layers.1.ffn_norm1.weight": "diffusion_pytorch_model-00001-of-00002.safetensors",
55
+ "layers.1.ffn_norm2.weight": "diffusion_pytorch_model-00001-of-00002.safetensors",
56
+ "layers.10.adaLN_modulation.1.bias": "diffusion_pytorch_model-00001-of-00002.safetensors",
57
+ "layers.10.adaLN_modulation.1.weight": "diffusion_pytorch_model-00001-of-00002.safetensors",
58
+ "layers.10.attention.gate": "diffusion_pytorch_model-00001-of-00002.safetensors",
59
+ "layers.10.attention.k_norm.bias": "diffusion_pytorch_model-00001-of-00002.safetensors",
60
+ "layers.10.attention.k_norm.weight": "diffusion_pytorch_model-00001-of-00002.safetensors",
61
+ "layers.10.attention.ky_norm.bias": "diffusion_pytorch_model-00001-of-00002.safetensors",
62
+ "layers.10.attention.ky_norm.weight": "diffusion_pytorch_model-00001-of-00002.safetensors",
63
+ "layers.10.attention.q_norm.bias": "diffusion_pytorch_model-00001-of-00002.safetensors",
64
+ "layers.10.attention.q_norm.weight": "diffusion_pytorch_model-00001-of-00002.safetensors",
65
+ "layers.10.attention.wk.weight": "diffusion_pytorch_model-00001-of-00002.safetensors",
66
+ "layers.10.attention.wk_y.weight": "diffusion_pytorch_model-00001-of-00002.safetensors",
67
+ "layers.10.attention.wo.weight": "diffusion_pytorch_model-00001-of-00002.safetensors",
68
+ "layers.10.attention.wq.weight": "diffusion_pytorch_model-00001-of-00002.safetensors",
69
+ "layers.10.attention.wv.weight": "diffusion_pytorch_model-00001-of-00002.safetensors",
70
+ "layers.10.attention.wv_y.weight": "diffusion_pytorch_model-00001-of-00002.safetensors",
71
+ "layers.10.attention_norm1.weight": "diffusion_pytorch_model-00001-of-00002.safetensors",
72
+ "layers.10.attention_norm2.weight": "diffusion_pytorch_model-00001-of-00002.safetensors",
73
+ "layers.10.attention_y_norm.weight": "diffusion_pytorch_model-00001-of-00002.safetensors",
74
+ "layers.10.feed_forward.w1.weight": "diffusion_pytorch_model-00001-of-00002.safetensors",
75
+ "layers.10.feed_forward.w2.weight": "diffusion_pytorch_model-00001-of-00002.safetensors",
76
+ "layers.10.feed_forward.w3.weight": "diffusion_pytorch_model-00001-of-00002.safetensors",
77
+ "layers.10.ffn_norm1.weight": "diffusion_pytorch_model-00001-of-00002.safetensors",
78
+ "layers.10.ffn_norm2.weight": "diffusion_pytorch_model-00001-of-00002.safetensors",
79
+ "layers.11.adaLN_modulation.1.bias": "diffusion_pytorch_model-00001-of-00002.safetensors",
80
+ "layers.11.adaLN_modulation.1.weight": "diffusion_pytorch_model-00001-of-00002.safetensors",
81
+ "layers.11.attention.gate": "diffusion_pytorch_model-00001-of-00002.safetensors",
82
+ "layers.11.attention.k_norm.bias": "diffusion_pytorch_model-00001-of-00002.safetensors",
83
+ "layers.11.attention.k_norm.weight": "diffusion_pytorch_model-00001-of-00002.safetensors",
84
+ "layers.11.attention.ky_norm.bias": "diffusion_pytorch_model-00001-of-00002.safetensors",
85
+ "layers.11.attention.ky_norm.weight": "diffusion_pytorch_model-00001-of-00002.safetensors",
86
+ "layers.11.attention.q_norm.bias": "diffusion_pytorch_model-00001-of-00002.safetensors",
87
+ "layers.11.attention.q_norm.weight": "diffusion_pytorch_model-00001-of-00002.safetensors",
88
+ "layers.11.attention.wk.weight": "diffusion_pytorch_model-00001-of-00002.safetensors",
89
+ "layers.11.attention.wk_y.weight": "diffusion_pytorch_model-00001-of-00002.safetensors",
90
+ "layers.11.attention.wo.weight": "diffusion_pytorch_model-00001-of-00002.safetensors",
91
+ "layers.11.attention.wq.weight": "diffusion_pytorch_model-00001-of-00002.safetensors",
92
+ "layers.11.attention.wv.weight": "diffusion_pytorch_model-00001-of-00002.safetensors",
93
+ "layers.11.attention.wv_y.weight": "diffusion_pytorch_model-00001-of-00002.safetensors",
94
+ "layers.11.attention_norm1.weight": "diffusion_pytorch_model-00001-of-00002.safetensors",
95
+ "layers.11.attention_norm2.weight": "diffusion_pytorch_model-00001-of-00002.safetensors",
96
+ "layers.11.attention_y_norm.weight": "diffusion_pytorch_model-00001-of-00002.safetensors",
97
+ "layers.11.feed_forward.w1.weight": "diffusion_pytorch_model-00001-of-00002.safetensors",
98
+ "layers.11.feed_forward.w2.weight": "diffusion_pytorch_model-00001-of-00002.safetensors",
99
+ "layers.11.feed_forward.w3.weight": "diffusion_pytorch_model-00001-of-00002.safetensors",
100
+ "layers.11.ffn_norm1.weight": "diffusion_pytorch_model-00001-of-00002.safetensors",
101
+ "layers.11.ffn_norm2.weight": "diffusion_pytorch_model-00001-of-00002.safetensors",
102
+ "layers.12.adaLN_modulation.1.bias": "diffusion_pytorch_model-00001-of-00002.safetensors",
103
+ "layers.12.adaLN_modulation.1.weight": "diffusion_pytorch_model-00001-of-00002.safetensors",
104
+ "layers.12.attention.gate": "diffusion_pytorch_model-00001-of-00002.safetensors",
105
+ "layers.12.attention.k_norm.bias": "diffusion_pytorch_model-00001-of-00002.safetensors",
106
+ "layers.12.attention.k_norm.weight": "diffusion_pytorch_model-00001-of-00002.safetensors",
107
+ "layers.12.attention.ky_norm.bias": "diffusion_pytorch_model-00001-of-00002.safetensors",
108
+ "layers.12.attention.ky_norm.weight": "diffusion_pytorch_model-00001-of-00002.safetensors",
109
+ "layers.12.attention.q_norm.bias": "diffusion_pytorch_model-00001-of-00002.safetensors",
110
+ "layers.12.attention.q_norm.weight": "diffusion_pytorch_model-00001-of-00002.safetensors",
111
+ "layers.12.attention.wk.weight": "diffusion_pytorch_model-00001-of-00002.safetensors",
112
+ "layers.12.attention.wk_y.weight": "diffusion_pytorch_model-00001-of-00002.safetensors",
113
+ "layers.12.attention.wo.weight": "diffusion_pytorch_model-00001-of-00002.safetensors",
114
+ "layers.12.attention.wq.weight": "diffusion_pytorch_model-00001-of-00002.safetensors",
115
+ "layers.12.attention.wv.weight": "diffusion_pytorch_model-00001-of-00002.safetensors",
116
+ "layers.12.attention.wv_y.weight": "diffusion_pytorch_model-00001-of-00002.safetensors",
117
+ "layers.12.attention_norm1.weight": "diffusion_pytorch_model-00001-of-00002.safetensors",
118
+ "layers.12.attention_norm2.weight": "diffusion_pytorch_model-00001-of-00002.safetensors",
119
+ "layers.12.attention_y_norm.weight": "diffusion_pytorch_model-00001-of-00002.safetensors",
120
+ "layers.12.feed_forward.w1.weight": "diffusion_pytorch_model-00001-of-00002.safetensors",
121
+ "layers.12.feed_forward.w2.weight": "diffusion_pytorch_model-00001-of-00002.safetensors",
122
+ "layers.12.feed_forward.w3.weight": "diffusion_pytorch_model-00001-of-00002.safetensors",
123
+ "layers.12.ffn_norm1.weight": "diffusion_pytorch_model-00001-of-00002.safetensors",
124
+ "layers.12.ffn_norm2.weight": "diffusion_pytorch_model-00001-of-00002.safetensors",
125
+ "layers.13.adaLN_modulation.1.bias": "diffusion_pytorch_model-00001-of-00002.safetensors",
126
+ "layers.13.adaLN_modulation.1.weight": "diffusion_pytorch_model-00001-of-00002.safetensors",
127
+ "layers.13.attention.gate": "diffusion_pytorch_model-00001-of-00002.safetensors",
128
+ "layers.13.attention.k_norm.bias": "diffusion_pytorch_model-00001-of-00002.safetensors",
129
+ "layers.13.attention.k_norm.weight": "diffusion_pytorch_model-00001-of-00002.safetensors",
130
+ "layers.13.attention.ky_norm.bias": "diffusion_pytorch_model-00001-of-00002.safetensors",
131
+ "layers.13.attention.ky_norm.weight": "diffusion_pytorch_model-00001-of-00002.safetensors",
132
+ "layers.13.attention.q_norm.bias": "diffusion_pytorch_model-00001-of-00002.safetensors",
133
+ "layers.13.attention.q_norm.weight": "diffusion_pytorch_model-00001-of-00002.safetensors",
134
+ "layers.13.attention.wk.weight": "diffusion_pytorch_model-00001-of-00002.safetensors",
135
+ "layers.13.attention.wk_y.weight": "diffusion_pytorch_model-00001-of-00002.safetensors",
136
+ "layers.13.attention.wo.weight": "diffusion_pytorch_model-00001-of-00002.safetensors",
137
+ "layers.13.attention.wq.weight": "diffusion_pytorch_model-00001-of-00002.safetensors",
138
+ "layers.13.attention.wv.weight": "diffusion_pytorch_model-00001-of-00002.safetensors",
139
+ "layers.13.attention.wv_y.weight": "diffusion_pytorch_model-00001-of-00002.safetensors",
140
+ "layers.13.attention_norm1.weight": "diffusion_pytorch_model-00001-of-00002.safetensors",
141
+ "layers.13.attention_norm2.weight": "diffusion_pytorch_model-00001-of-00002.safetensors",
142
+ "layers.13.attention_y_norm.weight": "diffusion_pytorch_model-00001-of-00002.safetensors",
143
+ "layers.13.feed_forward.w1.weight": "diffusion_pytorch_model-00001-of-00002.safetensors",
144
+ "layers.13.feed_forward.w2.weight": "diffusion_pytorch_model-00001-of-00002.safetensors",
145
+ "layers.13.feed_forward.w3.weight": "diffusion_pytorch_model-00001-of-00002.safetensors",
146
+ "layers.13.ffn_norm1.weight": "diffusion_pytorch_model-00001-of-00002.safetensors",
147
+ "layers.13.ffn_norm2.weight": "diffusion_pytorch_model-00001-of-00002.safetensors",
148
+ "layers.14.adaLN_modulation.1.bias": "diffusion_pytorch_model-00002-of-00002.safetensors",
149
+ "layers.14.adaLN_modulation.1.weight": "diffusion_pytorch_model-00002-of-00002.safetensors",
150
+ "layers.14.attention.gate": "diffusion_pytorch_model-00001-of-00002.safetensors",
151
+ "layers.14.attention.k_norm.bias": "diffusion_pytorch_model-00001-of-00002.safetensors",
152
+ "layers.14.attention.k_norm.weight": "diffusion_pytorch_model-00001-of-00002.safetensors",
153
+ "layers.14.attention.ky_norm.bias": "diffusion_pytorch_model-00001-of-00002.safetensors",
154
+ "layers.14.attention.ky_norm.weight": "diffusion_pytorch_model-00001-of-00002.safetensors",
155
+ "layers.14.attention.q_norm.bias": "diffusion_pytorch_model-00001-of-00002.safetensors",
156
+ "layers.14.attention.q_norm.weight": "diffusion_pytorch_model-00001-of-00002.safetensors",
157
+ "layers.14.attention.wk.weight": "diffusion_pytorch_model-00001-of-00002.safetensors",
158
+ "layers.14.attention.wk_y.weight": "diffusion_pytorch_model-00001-of-00002.safetensors",
159
+ "layers.14.attention.wo.weight": "diffusion_pytorch_model-00001-of-00002.safetensors",
160
+ "layers.14.attention.wq.weight": "diffusion_pytorch_model-00001-of-00002.safetensors",
161
+ "layers.14.attention.wv.weight": "diffusion_pytorch_model-00001-of-00002.safetensors",
162
+ "layers.14.attention.wv_y.weight": "diffusion_pytorch_model-00001-of-00002.safetensors",
163
+ "layers.14.attention_norm1.weight": "diffusion_pytorch_model-00002-of-00002.safetensors",
164
+ "layers.14.attention_norm2.weight": "diffusion_pytorch_model-00002-of-00002.safetensors",
165
+ "layers.14.attention_y_norm.weight": "diffusion_pytorch_model-00002-of-00002.safetensors",
166
+ "layers.14.feed_forward.w1.weight": "diffusion_pytorch_model-00002-of-00002.safetensors",
167
+ "layers.14.feed_forward.w2.weight": "diffusion_pytorch_model-00002-of-00002.safetensors",
168
+ "layers.14.feed_forward.w3.weight": "diffusion_pytorch_model-00002-of-00002.safetensors",
169
+ "layers.14.ffn_norm1.weight": "diffusion_pytorch_model-00002-of-00002.safetensors",
170
+ "layers.14.ffn_norm2.weight": "diffusion_pytorch_model-00002-of-00002.safetensors",
171
+ "layers.15.adaLN_modulation.1.bias": "diffusion_pytorch_model-00002-of-00002.safetensors",
172
+ "layers.15.adaLN_modulation.1.weight": "diffusion_pytorch_model-00002-of-00002.safetensors",
173
+ "layers.15.attention.gate": "diffusion_pytorch_model-00002-of-00002.safetensors",
174
+ "layers.15.attention.k_norm.bias": "diffusion_pytorch_model-00002-of-00002.safetensors",
175
+ "layers.15.attention.k_norm.weight": "diffusion_pytorch_model-00002-of-00002.safetensors",
176
+ "layers.15.attention.ky_norm.bias": "diffusion_pytorch_model-00002-of-00002.safetensors",
177
+ "layers.15.attention.ky_norm.weight": "diffusion_pytorch_model-00002-of-00002.safetensors",
178
+ "layers.15.attention.q_norm.bias": "diffusion_pytorch_model-00002-of-00002.safetensors",
179
+ "layers.15.attention.q_norm.weight": "diffusion_pytorch_model-00002-of-00002.safetensors",
180
+ "layers.15.attention.wk.weight": "diffusion_pytorch_model-00002-of-00002.safetensors",
181
+ "layers.15.attention.wk_y.weight": "diffusion_pytorch_model-00002-of-00002.safetensors",
182
+ "layers.15.attention.wo.weight": "diffusion_pytorch_model-00002-of-00002.safetensors",
183
+ "layers.15.attention.wq.weight": "diffusion_pytorch_model-00002-of-00002.safetensors",
184
+ "layers.15.attention.wv.weight": "diffusion_pytorch_model-00002-of-00002.safetensors",
185
+ "layers.15.attention.wv_y.weight": "diffusion_pytorch_model-00002-of-00002.safetensors",
186
+ "layers.15.attention_norm1.weight": "diffusion_pytorch_model-00002-of-00002.safetensors",
187
+ "layers.15.attention_norm2.weight": "diffusion_pytorch_model-00002-of-00002.safetensors",
188
+ "layers.15.attention_y_norm.weight": "diffusion_pytorch_model-00002-of-00002.safetensors",
189
+ "layers.15.feed_forward.w1.weight": "diffusion_pytorch_model-00002-of-00002.safetensors",
190
+ "layers.15.feed_forward.w2.weight": "diffusion_pytorch_model-00002-of-00002.safetensors",
191
+ "layers.15.feed_forward.w3.weight": "diffusion_pytorch_model-00002-of-00002.safetensors",
192
+ "layers.15.ffn_norm1.weight": "diffusion_pytorch_model-00002-of-00002.safetensors",
193
+ "layers.15.ffn_norm2.weight": "diffusion_pytorch_model-00002-of-00002.safetensors",
194
+ "layers.2.adaLN_modulation.1.bias": "diffusion_pytorch_model-00001-of-00002.safetensors",
195
+ "layers.2.adaLN_modulation.1.weight": "diffusion_pytorch_model-00001-of-00002.safetensors",
196
+ "layers.2.attention.gate": "diffusion_pytorch_model-00001-of-00002.safetensors",
197
+ "layers.2.attention.k_norm.bias": "diffusion_pytorch_model-00001-of-00002.safetensors",
198
+ "layers.2.attention.k_norm.weight": "diffusion_pytorch_model-00001-of-00002.safetensors",
199
+ "layers.2.attention.ky_norm.bias": "diffusion_pytorch_model-00001-of-00002.safetensors",
200
+ "layers.2.attention.ky_norm.weight": "diffusion_pytorch_model-00001-of-00002.safetensors",
201
+ "layers.2.attention.q_norm.bias": "diffusion_pytorch_model-00001-of-00002.safetensors",
202
+ "layers.2.attention.q_norm.weight": "diffusion_pytorch_model-00001-of-00002.safetensors",
203
+ "layers.2.attention.wk.weight": "diffusion_pytorch_model-00001-of-00002.safetensors",
204
+ "layers.2.attention.wk_y.weight": "diffusion_pytorch_model-00001-of-00002.safetensors",
205
+ "layers.2.attention.wo.weight": "diffusion_pytorch_model-00001-of-00002.safetensors",
206
+ "layers.2.attention.wq.weight": "diffusion_pytorch_model-00001-of-00002.safetensors",
207
+ "layers.2.attention.wv.weight": "diffusion_pytorch_model-00001-of-00002.safetensors",
208
+ "layers.2.attention.wv_y.weight": "diffusion_pytorch_model-00001-of-00002.safetensors",
209
+ "layers.2.attention_norm1.weight": "diffusion_pytorch_model-00001-of-00002.safetensors",
210
+ "layers.2.attention_norm2.weight": "diffusion_pytorch_model-00001-of-00002.safetensors",
211
+ "layers.2.attention_y_norm.weight": "diffusion_pytorch_model-00001-of-00002.safetensors",
212
+ "layers.2.feed_forward.w1.weight": "diffusion_pytorch_model-00001-of-00002.safetensors",
213
+ "layers.2.feed_forward.w2.weight": "diffusion_pytorch_model-00001-of-00002.safetensors",
214
+ "layers.2.feed_forward.w3.weight": "diffusion_pytorch_model-00001-of-00002.safetensors",
215
+ "layers.2.ffn_norm1.weight": "diffusion_pytorch_model-00001-of-00002.safetensors",
216
+ "layers.2.ffn_norm2.weight": "diffusion_pytorch_model-00001-of-00002.safetensors",
217
+ "layers.3.adaLN_modulation.1.bias": "diffusion_pytorch_model-00001-of-00002.safetensors",
218
+ "layers.3.adaLN_modulation.1.weight": "diffusion_pytorch_model-00001-of-00002.safetensors",
219
+ "layers.3.attention.gate": "diffusion_pytorch_model-00001-of-00002.safetensors",
220
+ "layers.3.attention.k_norm.bias": "diffusion_pytorch_model-00001-of-00002.safetensors",
221
+ "layers.3.attention.k_norm.weight": "diffusion_pytorch_model-00001-of-00002.safetensors",
222
+ "layers.3.attention.ky_norm.bias": "diffusion_pytorch_model-00001-of-00002.safetensors",
223
+ "layers.3.attention.ky_norm.weight": "diffusion_pytorch_model-00001-of-00002.safetensors",
224
+ "layers.3.attention.q_norm.bias": "diffusion_pytorch_model-00001-of-00002.safetensors",
225
+ "layers.3.attention.q_norm.weight": "diffusion_pytorch_model-00001-of-00002.safetensors",
226
+ "layers.3.attention.wk.weight": "diffusion_pytorch_model-00001-of-00002.safetensors",
227
+ "layers.3.attention.wk_y.weight": "diffusion_pytorch_model-00001-of-00002.safetensors",
228
+ "layers.3.attention.wo.weight": "diffusion_pytorch_model-00001-of-00002.safetensors",
229
+ "layers.3.attention.wq.weight": "diffusion_pytorch_model-00001-of-00002.safetensors",
230
+ "layers.3.attention.wv.weight": "diffusion_pytorch_model-00001-of-00002.safetensors",
231
+ "layers.3.attention.wv_y.weight": "diffusion_pytorch_model-00001-of-00002.safetensors",
232
+ "layers.3.attention_norm1.weight": "diffusion_pytorch_model-00001-of-00002.safetensors",
233
+ "layers.3.attention_norm2.weight": "diffusion_pytorch_model-00001-of-00002.safetensors",
234
+ "layers.3.attention_y_norm.weight": "diffusion_pytorch_model-00001-of-00002.safetensors",
235
+ "layers.3.feed_forward.w1.weight": "diffusion_pytorch_model-00001-of-00002.safetensors",
236
+ "layers.3.feed_forward.w2.weight": "diffusion_pytorch_model-00001-of-00002.safetensors",
237
+ "layers.3.feed_forward.w3.weight": "diffusion_pytorch_model-00001-of-00002.safetensors",
238
+ "layers.3.ffn_norm1.weight": "diffusion_pytorch_model-00001-of-00002.safetensors",
239
+ "layers.3.ffn_norm2.weight": "diffusion_pytorch_model-00001-of-00002.safetensors",
240
+ "layers.4.adaLN_modulation.1.bias": "diffusion_pytorch_model-00001-of-00002.safetensors",
241
+ "layers.4.adaLN_modulation.1.weight": "diffusion_pytorch_model-00001-of-00002.safetensors",
242
+ "layers.4.attention.gate": "diffusion_pytorch_model-00001-of-00002.safetensors",
243
+ "layers.4.attention.k_norm.bias": "diffusion_pytorch_model-00001-of-00002.safetensors",
244
+ "layers.4.attention.k_norm.weight": "diffusion_pytorch_model-00001-of-00002.safetensors",
245
+ "layers.4.attention.ky_norm.bias": "diffusion_pytorch_model-00001-of-00002.safetensors",
246
+ "layers.4.attention.ky_norm.weight": "diffusion_pytorch_model-00001-of-00002.safetensors",
247
+ "layers.4.attention.q_norm.bias": "diffusion_pytorch_model-00001-of-00002.safetensors",
248
+ "layers.4.attention.q_norm.weight": "diffusion_pytorch_model-00001-of-00002.safetensors",
249
+ "layers.4.attention.wk.weight": "diffusion_pytorch_model-00001-of-00002.safetensors",
250
+ "layers.4.attention.wk_y.weight": "diffusion_pytorch_model-00001-of-00002.safetensors",
251
+ "layers.4.attention.wo.weight": "diffusion_pytorch_model-00001-of-00002.safetensors",
252
+ "layers.4.attention.wq.weight": "diffusion_pytorch_model-00001-of-00002.safetensors",
253
+ "layers.4.attention.wv.weight": "diffusion_pytorch_model-00001-of-00002.safetensors",
254
+ "layers.4.attention.wv_y.weight": "diffusion_pytorch_model-00001-of-00002.safetensors",
255
+ "layers.4.attention_norm1.weight": "diffusion_pytorch_model-00001-of-00002.safetensors",
256
+ "layers.4.attention_norm2.weight": "diffusion_pytorch_model-00001-of-00002.safetensors",
257
+ "layers.4.attention_y_norm.weight": "diffusion_pytorch_model-00001-of-00002.safetensors",
258
+ "layers.4.feed_forward.w1.weight": "diffusion_pytorch_model-00001-of-00002.safetensors",
259
+ "layers.4.feed_forward.w2.weight": "diffusion_pytorch_model-00001-of-00002.safetensors",
260
+ "layers.4.feed_forward.w3.weight": "diffusion_pytorch_model-00001-of-00002.safetensors",
261
+ "layers.4.ffn_norm1.weight": "diffusion_pytorch_model-00001-of-00002.safetensors",
262
+ "layers.4.ffn_norm2.weight": "diffusion_pytorch_model-00001-of-00002.safetensors",
263
+ "layers.5.adaLN_modulation.1.bias": "diffusion_pytorch_model-00001-of-00002.safetensors",
264
+ "layers.5.adaLN_modulation.1.weight": "diffusion_pytorch_model-00001-of-00002.safetensors",
265
+ "layers.5.attention.gate": "diffusion_pytorch_model-00001-of-00002.safetensors",
266
+ "layers.5.attention.k_norm.bias": "diffusion_pytorch_model-00001-of-00002.safetensors",
267
+ "layers.5.attention.k_norm.weight": "diffusion_pytorch_model-00001-of-00002.safetensors",
268
+ "layers.5.attention.ky_norm.bias": "diffusion_pytorch_model-00001-of-00002.safetensors",
269
+ "layers.5.attention.ky_norm.weight": "diffusion_pytorch_model-00001-of-00002.safetensors",
270
+ "layers.5.attention.q_norm.bias": "diffusion_pytorch_model-00001-of-00002.safetensors",
271
+ "layers.5.attention.q_norm.weight": "diffusion_pytorch_model-00001-of-00002.safetensors",
272
+ "layers.5.attention.wk.weight": "diffusion_pytorch_model-00001-of-00002.safetensors",
273
+ "layers.5.attention.wk_y.weight": "diffusion_pytorch_model-00001-of-00002.safetensors",
274
+ "layers.5.attention.wo.weight": "diffusion_pytorch_model-00001-of-00002.safetensors",
275
+ "layers.5.attention.wq.weight": "diffusion_pytorch_model-00001-of-00002.safetensors",
276
+ "layers.5.attention.wv.weight": "diffusion_pytorch_model-00001-of-00002.safetensors",
277
+ "layers.5.attention.wv_y.weight": "diffusion_pytorch_model-00001-of-00002.safetensors",
278
+ "layers.5.attention_norm1.weight": "diffusion_pytorch_model-00001-of-00002.safetensors",
279
+ "layers.5.attention_norm2.weight": "diffusion_pytorch_model-00001-of-00002.safetensors",
280
+ "layers.5.attention_y_norm.weight": "diffusion_pytorch_model-00001-of-00002.safetensors",
281
+ "layers.5.feed_forward.w1.weight": "diffusion_pytorch_model-00001-of-00002.safetensors",
282
+ "layers.5.feed_forward.w2.weight": "diffusion_pytorch_model-00001-of-00002.safetensors",
283
+ "layers.5.feed_forward.w3.weight": "diffusion_pytorch_model-00001-of-00002.safetensors",
284
+ "layers.5.ffn_norm1.weight": "diffusion_pytorch_model-00001-of-00002.safetensors",
285
+ "layers.5.ffn_norm2.weight": "diffusion_pytorch_model-00001-of-00002.safetensors",
286
+ "layers.6.adaLN_modulation.1.bias": "diffusion_pytorch_model-00001-of-00002.safetensors",
287
+ "layers.6.adaLN_modulation.1.weight": "diffusion_pytorch_model-00001-of-00002.safetensors",
288
+ "layers.6.attention.gate": "diffusion_pytorch_model-00001-of-00002.safetensors",
289
+ "layers.6.attention.k_norm.bias": "diffusion_pytorch_model-00001-of-00002.safetensors",
290
+ "layers.6.attention.k_norm.weight": "diffusion_pytorch_model-00001-of-00002.safetensors",
291
+ "layers.6.attention.ky_norm.bias": "diffusion_pytorch_model-00001-of-00002.safetensors",
292
+ "layers.6.attention.ky_norm.weight": "diffusion_pytorch_model-00001-of-00002.safetensors",
293
+ "layers.6.attention.q_norm.bias": "diffusion_pytorch_model-00001-of-00002.safetensors",
294
+ "layers.6.attention.q_norm.weight": "diffusion_pytorch_model-00001-of-00002.safetensors",
295
+ "layers.6.attention.wk.weight": "diffusion_pytorch_model-00001-of-00002.safetensors",
296
+ "layers.6.attention.wk_y.weight": "diffusion_pytorch_model-00001-of-00002.safetensors",
297
+ "layers.6.attention.wo.weight": "diffusion_pytorch_model-00001-of-00002.safetensors",
298
+ "layers.6.attention.wq.weight": "diffusion_pytorch_model-00001-of-00002.safetensors",
299
+ "layers.6.attention.wv.weight": "diffusion_pytorch_model-00001-of-00002.safetensors",
300
+ "layers.6.attention.wv_y.weight": "diffusion_pytorch_model-00001-of-00002.safetensors",
301
+ "layers.6.attention_norm1.weight": "diffusion_pytorch_model-00001-of-00002.safetensors",
302
+ "layers.6.attention_norm2.weight": "diffusion_pytorch_model-00001-of-00002.safetensors",
303
+ "layers.6.attention_y_norm.weight": "diffusion_pytorch_model-00001-of-00002.safetensors",
304
+ "layers.6.feed_forward.w1.weight": "diffusion_pytorch_model-00001-of-00002.safetensors",
305
+ "layers.6.feed_forward.w2.weight": "diffusion_pytorch_model-00001-of-00002.safetensors",
306
+ "layers.6.feed_forward.w3.weight": "diffusion_pytorch_model-00001-of-00002.safetensors",
307
+ "layers.6.ffn_norm1.weight": "diffusion_pytorch_model-00001-of-00002.safetensors",
308
+ "layers.6.ffn_norm2.weight": "diffusion_pytorch_model-00001-of-00002.safetensors",
309
+ "layers.7.adaLN_modulation.1.bias": "diffusion_pytorch_model-00001-of-00002.safetensors",
310
+ "layers.7.adaLN_modulation.1.weight": "diffusion_pytorch_model-00001-of-00002.safetensors",
311
+ "layers.7.attention.gate": "diffusion_pytorch_model-00001-of-00002.safetensors",
312
+ "layers.7.attention.k_norm.bias": "diffusion_pytorch_model-00001-of-00002.safetensors",
313
+ "layers.7.attention.k_norm.weight": "diffusion_pytorch_model-00001-of-00002.safetensors",
314
+ "layers.7.attention.ky_norm.bias": "diffusion_pytorch_model-00001-of-00002.safetensors",
315
+ "layers.7.attention.ky_norm.weight": "diffusion_pytorch_model-00001-of-00002.safetensors",
316
+ "layers.7.attention.q_norm.bias": "diffusion_pytorch_model-00001-of-00002.safetensors",
317
+ "layers.7.attention.q_norm.weight": "diffusion_pytorch_model-00001-of-00002.safetensors",
318
+ "layers.7.attention.wk.weight": "diffusion_pytorch_model-00001-of-00002.safetensors",
319
+ "layers.7.attention.wk_y.weight": "diffusion_pytorch_model-00001-of-00002.safetensors",
320
+ "layers.7.attention.wo.weight": "diffusion_pytorch_model-00001-of-00002.safetensors",
321
+ "layers.7.attention.wq.weight": "diffusion_pytorch_model-00001-of-00002.safetensors",
322
+ "layers.7.attention.wv.weight": "diffusion_pytorch_model-00001-of-00002.safetensors",
323
+ "layers.7.attention.wv_y.weight": "diffusion_pytorch_model-00001-of-00002.safetensors",
324
+ "layers.7.attention_norm1.weight": "diffusion_pytorch_model-00001-of-00002.safetensors",
325
+ "layers.7.attention_norm2.weight": "diffusion_pytorch_model-00001-of-00002.safetensors",
326
+ "layers.7.attention_y_norm.weight": "diffusion_pytorch_model-00001-of-00002.safetensors",
327
+ "layers.7.feed_forward.w1.weight": "diffusion_pytorch_model-00001-of-00002.safetensors",
328
+ "layers.7.feed_forward.w2.weight": "diffusion_pytorch_model-00001-of-00002.safetensors",
329
+ "layers.7.feed_forward.w3.weight": "diffusion_pytorch_model-00001-of-00002.safetensors",
330
+ "layers.7.ffn_norm1.weight": "diffusion_pytorch_model-00001-of-00002.safetensors",
331
+ "layers.7.ffn_norm2.weight": "diffusion_pytorch_model-00001-of-00002.safetensors",
332
+ "layers.8.adaLN_modulation.1.bias": "diffusion_pytorch_model-00001-of-00002.safetensors",
333
+ "layers.8.adaLN_modulation.1.weight": "diffusion_pytorch_model-00001-of-00002.safetensors",
334
+ "layers.8.attention.gate": "diffusion_pytorch_model-00001-of-00002.safetensors",
335
+ "layers.8.attention.k_norm.bias": "diffusion_pytorch_model-00001-of-00002.safetensors",
336
+ "layers.8.attention.k_norm.weight": "diffusion_pytorch_model-00001-of-00002.safetensors",
337
+ "layers.8.attention.ky_norm.bias": "diffusion_pytorch_model-00001-of-00002.safetensors",
338
+ "layers.8.attention.ky_norm.weight": "diffusion_pytorch_model-00001-of-00002.safetensors",
339
+ "layers.8.attention.q_norm.bias": "diffusion_pytorch_model-00001-of-00002.safetensors",
340
+ "layers.8.attention.q_norm.weight": "diffusion_pytorch_model-00001-of-00002.safetensors",
341
+ "layers.8.attention.wk.weight": "diffusion_pytorch_model-00001-of-00002.safetensors",
342
+ "layers.8.attention.wk_y.weight": "diffusion_pytorch_model-00001-of-00002.safetensors",
343
+ "layers.8.attention.wo.weight": "diffusion_pytorch_model-00001-of-00002.safetensors",
344
+ "layers.8.attention.wq.weight": "diffusion_pytorch_model-00001-of-00002.safetensors",
345
+ "layers.8.attention.wv.weight": "diffusion_pytorch_model-00001-of-00002.safetensors",
346
+ "layers.8.attention.wv_y.weight": "diffusion_pytorch_model-00001-of-00002.safetensors",
347
+ "layers.8.attention_norm1.weight": "diffusion_pytorch_model-00001-of-00002.safetensors",
348
+ "layers.8.attention_norm2.weight": "diffusion_pytorch_model-00001-of-00002.safetensors",
349
+ "layers.8.attention_y_norm.weight": "diffusion_pytorch_model-00001-of-00002.safetensors",
350
+ "layers.8.feed_forward.w1.weight": "diffusion_pytorch_model-00001-of-00002.safetensors",
351
+ "layers.8.feed_forward.w2.weight": "diffusion_pytorch_model-00001-of-00002.safetensors",
352
+ "layers.8.feed_forward.w3.weight": "diffusion_pytorch_model-00001-of-00002.safetensors",
353
+ "layers.8.ffn_norm1.weight": "diffusion_pytorch_model-00001-of-00002.safetensors",
354
+ "layers.8.ffn_norm2.weight": "diffusion_pytorch_model-00001-of-00002.safetensors",
355
+ "layers.9.adaLN_modulation.1.bias": "diffusion_pytorch_model-00001-of-00002.safetensors",
356
+ "layers.9.adaLN_modulation.1.weight": "diffusion_pytorch_model-00001-of-00002.safetensors",
357
+ "layers.9.attention.gate": "diffusion_pytorch_model-00001-of-00002.safetensors",
358
+ "layers.9.attention.k_norm.bias": "diffusion_pytorch_model-00001-of-00002.safetensors",
359
+ "layers.9.attention.k_norm.weight": "diffusion_pytorch_model-00001-of-00002.safetensors",
360
+ "layers.9.attention.ky_norm.bias": "diffusion_pytorch_model-00001-of-00002.safetensors",
361
+ "layers.9.attention.ky_norm.weight": "diffusion_pytorch_model-00001-of-00002.safetensors",
362
+ "layers.9.attention.q_norm.bias": "diffusion_pytorch_model-00001-of-00002.safetensors",
363
+ "layers.9.attention.q_norm.weight": "diffusion_pytorch_model-00001-of-00002.safetensors",
364
+ "layers.9.attention.wk.weight": "diffusion_pytorch_model-00001-of-00002.safetensors",
365
+ "layers.9.attention.wk_y.weight": "diffusion_pytorch_model-00001-of-00002.safetensors",
366
+ "layers.9.attention.wo.weight": "diffusion_pytorch_model-00001-of-00002.safetensors",
367
+ "layers.9.attention.wq.weight": "diffusion_pytorch_model-00001-of-00002.safetensors",
368
+ "layers.9.attention.wv.weight": "diffusion_pytorch_model-00001-of-00002.safetensors",
369
+ "layers.9.attention.wv_y.weight": "diffusion_pytorch_model-00001-of-00002.safetensors",
370
+ "layers.9.attention_norm1.weight": "diffusion_pytorch_model-00001-of-00002.safetensors",
371
+ "layers.9.attention_norm2.weight": "diffusion_pytorch_model-00001-of-00002.safetensors",
372
+ "layers.9.attention_y_norm.weight": "diffusion_pytorch_model-00001-of-00002.safetensors",
373
+ "layers.9.feed_forward.w1.weight": "diffusion_pytorch_model-00001-of-00002.safetensors",
374
+ "layers.9.feed_forward.w2.weight": "diffusion_pytorch_model-00001-of-00002.safetensors",
375
+ "layers.9.feed_forward.w3.weight": "diffusion_pytorch_model-00001-of-00002.safetensors",
376
+ "layers.9.ffn_norm1.weight": "diffusion_pytorch_model-00001-of-00002.safetensors",
377
+ "layers.9.ffn_norm2.weight": "diffusion_pytorch_model-00001-of-00002.safetensors",
378
+ "t_embedder.mlp.0.bias": "diffusion_pytorch_model-00001-of-00002.safetensors",
379
+ "t_embedder.mlp.0.weight": "diffusion_pytorch_model-00001-of-00002.safetensors",
380
+ "t_embedder.mlp.2.bias": "diffusion_pytorch_model-00001-of-00002.safetensors",
381
+ "t_embedder.mlp.2.weight": "diffusion_pytorch_model-00001-of-00002.safetensors",
382
+ "x_embedder.bias": "diffusion_pytorch_model-00001-of-00002.safetensors",
383
+ "x_embedder.weight": "diffusion_pytorch_model-00001-of-00002.safetensors",
384
+ "y_embedder.0.bias": "diffusion_pytorch_model-00001-of-00002.safetensors",
385
+ "y_embedder.0.weight": "diffusion_pytorch_model-00001-of-00002.safetensors",
386
+ "y_embedder.1.bias": "diffusion_pytorch_model-00001-of-00002.safetensors",
387
+ "y_embedder.1.weight": "diffusion_pytorch_model-00001-of-00002.safetensors"
388
+ }
389
+ }
vae/config.json ADDED
@@ -0,0 +1,36 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "_class_name": "AutoencoderKL",
3
+ "_diffusers_version": "0.29.0.dev0",
4
+ "act_fn": "silu",
5
+ "block_out_channels": [
6
+ 128,
7
+ 256,
8
+ 512,
9
+ 512
10
+ ],
11
+ "down_block_types": [
12
+ "DownEncoderBlock2D",
13
+ "DownEncoderBlock2D",
14
+ "DownEncoderBlock2D",
15
+ "DownEncoderBlock2D"
16
+ ],
17
+ "force_upcast": true,
18
+ "in_channels": 3,
19
+ "latent_channels": 16,
20
+ "latents_mean": null,
21
+ "latents_std": null,
22
+ "layers_per_block": 2,
23
+ "norm_num_groups": 32,
24
+ "out_channels": 3,
25
+ "sample_size": 1024,
26
+ "scaling_factor": 1.5305,
27
+ "shift_factor": 0.0609,
28
+ "up_block_types": [
29
+ "UpDecoderBlock2D",
30
+ "UpDecoderBlock2D",
31
+ "UpDecoderBlock2D",
32
+ "UpDecoderBlock2D"
33
+ ],
34
+ "use_post_quant_conv": false,
35
+ "use_quant_conv": false
36
+ }
vae/diffusion_pytorch_model.fp16.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:f9b67a279283625caee39d61eacb5324243848477b4eb535355eaaa8423d4e09
3
+ size 167666654
vae/diffusion_pytorch_model.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:f9b67a279283625caee39d61eacb5324243848477b4eb535355eaaa8423d4e09
3
+ size 167666654