Spaces:
Sleeping
Sleeping
{ | |
"4": { | |
"inputs": { | |
"ckpt_name": "sd_xl_base_1.0.safetensors" | |
}, | |
"class_type": "CheckpointLoaderSimple" | |
}, | |
"5": { | |
"inputs": { | |
"width": 1024, | |
"height": 1024, | |
"batch_size": 1 | |
}, | |
"class_type": "EmptyLatentImage" | |
}, | |
"6": { | |
"inputs": { | |
"text": "a photo of a cat", | |
"clip": [ | |
"4", | |
1 | |
] | |
}, | |
"class_type": "CLIPTextEncode" | |
}, | |
"10": { | |
"inputs": { | |
"add_noise": "enable", | |
"noise_seed": 42, | |
"steps": 20, | |
"cfg": 7.5, | |
"sampler_name": "euler", | |
"scheduler": "normal", | |
"start_at_step": 0, | |
"end_at_step": 32, | |
"return_with_leftover_noise": "enable", | |
"model": [ | |
"4", | |
0 | |
], | |
"positive": [ | |
"6", | |
0 | |
], | |
"negative": [ | |
"15", | |
0 | |
], | |
"latent_image": [ | |
"5", | |
0 | |
] | |
}, | |
"class_type": "KSamplerAdvanced" | |
}, | |
"12": { | |
"inputs": { | |
"samples": [ | |
"14", | |
0 | |
], | |
"vae": [ | |
"4", | |
2 | |
] | |
}, | |
"class_type": "VAEDecode" | |
}, | |
"13": { | |
"inputs": { | |
"filename_prefix": "test_inference", | |
"images": [ | |
"12", | |
0 | |
] | |
}, | |
"class_type": "SaveImage" | |
}, | |
"14": { | |
"inputs": { | |
"add_noise": "disable", | |
"noise_seed": 42, | |
"steps": 20, | |
"cfg": 7.5, | |
"sampler_name": "euler", | |
"scheduler": "normal", | |
"start_at_step": 32, | |
"end_at_step": 10000, | |
"return_with_leftover_noise": "disable", | |
"model": [ | |
"16", | |
0 | |
], | |
"positive": [ | |
"17", | |
0 | |
], | |
"negative": [ | |
"20", | |
0 | |
], | |
"latent_image": [ | |
"10", | |
0 | |
] | |
}, | |
"class_type": "KSamplerAdvanced" | |
}, | |
"15": { | |
"inputs": { | |
"conditioning": [ | |
"6", | |
0 | |
] | |
}, | |
"class_type": "ConditioningZeroOut" | |
}, | |
"16": { | |
"inputs": { | |
"ckpt_name": "sd_xl_refiner_1.0.safetensors" | |
}, | |
"class_type": "CheckpointLoaderSimple" | |
}, | |
"17": { | |
"inputs": { | |
"text": "a photo of a cat", | |
"clip": [ | |
"16", | |
1 | |
] | |
}, | |
"class_type": "CLIPTextEncode" | |
}, | |
"20": { | |
"inputs": { | |
"text": "", | |
"clip": [ | |
"16", | |
1 | |
] | |
}, | |
"class_type": "CLIPTextEncode" | |
} | |
} |