frutiemax commited on
Commit
c4329f9
1 Parent(s): 25aa809

Upload PixArtSigmaPipeline

Browse files
README.md CHANGED
@@ -1 +1,4 @@
 
 
 
1
  This is still not trained, this is an attempt at sticking a 16 channels VAE to a Pixart-Sigma pipeline.
 
1
+ ---
2
+ library_name: diffusers
3
+ ---
4
  This is still not trained, this is an attempt at sticking a 16 channels VAE to a Pixart-Sigma pipeline.
transformer/config.json CHANGED
@@ -1,14 +1,16 @@
1
  {
2
  "_class_name": "PixArtTransformer2DModel",
3
  "_diffusers_version": "0.30.2",
 
4
  "activation_fn": "gelu-approximate",
5
  "attention_bias": true,
6
  "attention_head_dim": 72,
7
  "attention_type": "default",
8
  "caption_channels": 4096,
9
  "cross_attention_dim": 1152,
 
10
  "dropout": 0.0,
11
- "in_channels": 16,
12
  "interpolation_scale": 1,
13
  "norm_elementwise_affine": false,
14
  "norm_eps": 1e-06,
@@ -17,9 +19,12 @@
17
  "num_attention_heads": 16,
18
  "num_embeds_ada_norm": 1000,
19
  "num_layers": 28,
20
- "out_channels": 32,
 
 
21
  "patch_size": 2,
22
  "sample_size": 64,
23
  "upcast_attention": false,
24
- "use_additional_conditions": false
 
25
  }
 
1
  {
2
  "_class_name": "PixArtTransformer2DModel",
3
  "_diffusers_version": "0.30.2",
4
+ "_name_or_path": "frutiemax/TwistedReality-pixart-512ms",
5
  "activation_fn": "gelu-approximate",
6
  "attention_bias": true,
7
  "attention_head_dim": 72,
8
  "attention_type": "default",
9
  "caption_channels": 4096,
10
  "cross_attention_dim": 1152,
11
+ "double_self_attention": false,
12
  "dropout": 0.0,
13
+ "in_channels": 4,
14
  "interpolation_scale": 1,
15
  "norm_elementwise_affine": false,
16
  "norm_eps": 1e-06,
 
19
  "num_attention_heads": 16,
20
  "num_embeds_ada_norm": 1000,
21
  "num_layers": 28,
22
+ "num_vector_embeds": null,
23
+ "only_cross_attention": false,
24
+ "out_channels": 8,
25
  "patch_size": 2,
26
  "sample_size": 64,
27
  "upcast_attention": false,
28
+ "use_additional_conditions": null,
29
+ "use_linear_projection": false
30
  }
transformer/diffusion_pytorch_model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:d4a16d9e4b6c2b1c89a204e1435d3bddfe05d93fc7bca9dd6d29a408331bc6b9
3
- size 2444156432
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:a92c9ace52b60fe1da0c5997859c8ecc6ad1da55a68d5519110c89b66b59ee61
3
+ size 2443492488
vae/config.json CHANGED
@@ -1,12 +1,12 @@
1
  {
2
  "_class_name": "AutoencoderKL",
3
  "_diffusers_version": "0.30.2",
4
- "_name_or_path": "AuraDiffusion/16ch-vae",
5
  "act_fn": "silu",
6
  "block_out_channels": [
 
7
  128,
8
  256,
9
- 512,
10
  512
11
  ],
12
  "down_block_types": [
@@ -15,18 +15,18 @@
15
  "DownEncoderBlock2D",
16
  "DownEncoderBlock2D"
17
  ],
18
- "force_upcast": false,
19
  "in_channels": 3,
20
  "latent_channels": 16,
21
  "latents_mean": null,
22
  "latents_std": null,
23
  "layers_per_block": 2,
24
- "mid_block_add_attention": false,
25
  "norm_num_groups": 32,
26
  "out_channels": 3,
27
- "sample_size": 1024,
28
- "scaling_factor": 0.18215,
29
- "shift_factor": 0,
30
  "up_block_types": [
31
  "UpDecoderBlock2D",
32
  "UpDecoderBlock2D",
 
1
  {
2
  "_class_name": "AutoencoderKL",
3
  "_diffusers_version": "0.30.2",
4
+ "_name_or_path": "ostris/vae-kl-f8-d16",
5
  "act_fn": "silu",
6
  "block_out_channels": [
7
+ 128,
8
  128,
9
  256,
 
10
  512
11
  ],
12
  "down_block_types": [
 
15
  "DownEncoderBlock2D",
16
  "DownEncoderBlock2D"
17
  ],
18
+ "force_upcast": true,
19
  "in_channels": 3,
20
  "latent_channels": 16,
21
  "latents_mean": null,
22
  "latents_std": null,
23
  "layers_per_block": 2,
24
+ "mid_block_add_attention": true,
25
  "norm_num_groups": 32,
26
  "out_channels": 3,
27
+ "sample_size": 512,
28
+ "scaling_factor": 0.2614,
29
+ "shift_factor": null,
30
  "up_block_types": [
31
  "UpDecoderBlock2D",
32
  "UpDecoderBlock2D",
vae/diffusion_pytorch_model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:ead97b12214d10d347a57425f2387242710c8e101e09466bcf91bc0f28d65e9c
3
- size 326896452
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:b6468f242218bc72981284b6fda570cc7c07ce043e84979cc1bb7d99d39a64a2
3
+ size 229094420