{ "last_node_id": 18, "last_link_id": 32, "nodes": [ { "id": 4, "type": "CheckpointLoaderSimple", "pos": [ 50, 730 ], "size": { "0": 315, "1": 98 }, "flags": {}, "order": 0, "mode": 0, "outputs": [ { "name": "MODEL", "type": "MODEL", "links": [ 29 ], "slot_index": 0 }, { "name": "CLIP", "type": "CLIP", "links": [ 3, 5 ], "slot_index": 1 }, { "name": "VAE", "type": "VAE", "links": [ 8 ], "slot_index": 2 } ], "properties": { "Node name for S&R": "CheckpointLoaderSimple" }, "widgets_values": [ "sd15/realisticVisionV51_v51VAE.safetensors" ] }, { "id": 7, "type": "CLIPTextEncode", "pos": [ 690, 840 ], "size": { "0": 425.27801513671875, "1": 180.6060791015625 }, "flags": {}, "order": 6, "mode": 0, "inputs": [ { "name": "clip", "type": "CLIP", "link": 5 } ], "outputs": [ { "name": "CONDITIONING", "type": "CONDITIONING", "links": [ 6 ], "slot_index": 0 } ], "properties": { "Node name for S&R": "CLIPTextEncode" }, "widgets_values": [ "blurry, noisy, messy, lowres, jpeg, artifacts, ill, distorted, malformed" ] }, { "id": 8, "type": "VAEDecode", "pos": [ 1570, 700 ], "size": { "0": 140, "1": 46 }, "flags": {}, "order": 9, "mode": 0, "inputs": [ { "name": "samples", "type": "LATENT", "link": 7 }, { "name": "vae", "type": "VAE", "link": 8 } ], "outputs": [ { "name": "IMAGE", "type": "IMAGE", "links": [ 9 ], "slot_index": 0 } ], "properties": { "Node name for S&R": "VAEDecode" } }, { "id": 12, "type": "LoadImage", "pos": [ 250, 290 ], "size": { "0": 315, "1": 314 }, "flags": {}, "order": 1, "mode": 0, "outputs": [ { "name": "IMAGE", "type": "IMAGE", "links": [ 27 ], "shape": 3, "slot_index": 0 }, { "name": "MASK", "type": "MASK", "links": null, "shape": 3 } ], "properties": { "Node name for S&R": "LoadImage" }, "widgets_values": [ "girl_sitting.png", "image" ] }, { "id": 6, "type": "CLIPTextEncode", "pos": [ 690, 610 ], "size": { "0": 422.84503173828125, "1": 164.31304931640625 }, "flags": {}, "order": 5, "mode": 0, "inputs": [ { "name": "clip", "type": "CLIP", "link": 3 } ], "outputs": [ { "name": "CONDITIONING", "type": "CONDITIONING", "links": [ 4 ], "slot_index": 0 } ], "properties": { "Node name for S&R": "CLIPTextEncode" }, "widgets_values": [ "in a peaceful spring morning a woman wearing a white shirt is sitting in a park on a bench\n\nhigh quality, detailed, diffuse light" ] }, { "id": 5, "type": "EmptyLatentImage", "pos": [ 801, 1097 ], "size": { "0": 315, "1": 106 }, "flags": {}, "order": 2, "mode": 0, "outputs": [ { "name": "LATENT", "type": "LATENT", "links": [ 2 ], "slot_index": 0 } ], "properties": { "Node name for S&R": "EmptyLatentImage" }, "widgets_values": [ 512, 768, 1 ] }, { "id": 3, "type": "KSampler", "pos": [ 1210, 700 ], "size": { "0": 315, "1": 262 }, "flags": {}, "order": 8, "mode": 0, "inputs": [ { "name": "model", "type": "MODEL", "link": 30 }, { "name": "positive", "type": "CONDITIONING", "link": 4 }, { "name": "negative", "type": "CONDITIONING", "link": 6 }, { "name": "latent_image", "type": "LATENT", "link": 2 } ], "outputs": [ { "name": "LATENT", "type": "LATENT", "links": [ 7 ], "slot_index": 0 } ], "properties": { "Node name for S&R": "KSampler" }, "widgets_values": [ 2, "fixed", 30, 6.5, "ddpm", "karras", 1 ] }, { "id": 18, "type": "IPAdapterTiled", "pos": [ 700, 230 ], "size": { "0": 315, "1": 302 }, "flags": {}, "order": 7, "mode": 0, "inputs": [ { "name": "model", "type": "MODEL", "link": 29 }, { "name": "ipadapter", "type": "IPADAPTER", "link": 31 }, { "name": "image", "type": "IMAGE", "link": 27 }, { "name": "image_negative", "type": "IMAGE", "link": null }, { "name": "attn_mask", "type": "MASK", "link": null }, { "name": "clip_vision", "type": "CLIP_VISION", "link": 32 } ], "outputs": [ { "name": "MODEL", "type": "MODEL", "links": [ 30 ], "shape": 3, "slot_index": 0 }, { "name": "tiles", "type": "IMAGE", "links": null, "shape": 3 }, { "name": "masks", "type": "MASK", "links": null, "shape": 3 } ], "properties": { "Node name for S&R": "IPAdapterTiled" }, "widgets_values": [ 0.7000000000000001, "ease in", "concat", 0, 1, 0, "V only" ] }, { "id": 9, "type": "SaveImage", "pos": [ 1768, 700 ], "size": { "0": 529.7760009765625, "1": 582.3048095703125 }, "flags": {}, "order": 10, "mode": 0, "inputs": [ { "name": "images", "type": "IMAGE", "link": 9 } ], "properties": {}, "widgets_values": [ "IPAdapter" ] }, { "id": 15, "type": "IPAdapterModelLoader", "pos": [ 250, 70 ], "size": { "0": 315, "1": 58 }, "flags": {}, "order": 3, "mode": 0, "outputs": [ { "name": "IPADAPTER", "type": "IPADAPTER", "links": [ 31 ], "shape": 3, "slot_index": 0 } ], "properties": { "Node name for S&R": "IPAdapterModelLoader" }, "widgets_values": [ "ip-adapter-plus_sd15.safetensors" ] }, { "id": 16, "type": "CLIPVisionLoader", "pos": [ 250, 180 ], "size": { "0": 315, "1": 58 }, "flags": {}, "order": 4, "mode": 0, "outputs": [ { "name": "CLIP_VISION", "type": "CLIP_VISION", "links": [ 32 ], "shape": 3, "slot_index": 0 } ], "properties": { "Node name for S&R": "CLIPVisionLoader" }, "widgets_values": [ "CLIP-ViT-H-14-laion2B-s32B-b79K.safetensors" ] } ], "links": [ [ 2, 5, 0, 3, 3, "LATENT" ], [ 3, 4, 1, 6, 0, "CLIP" ], [ 4, 6, 0, 3, 1, "CONDITIONING" ], [ 5, 4, 1, 7, 0, "CLIP" ], [ 6, 7, 0, 3, 2, "CONDITIONING" ], [ 7, 3, 0, 8, 0, "LATENT" ], [ 8, 4, 2, 8, 1, "VAE" ], [ 9, 8, 0, 9, 0, "IMAGE" ], [ 27, 12, 0, 18, 2, "IMAGE" ], [ 29, 4, 0, 18, 0, "MODEL" ], [ 30, 18, 0, 3, 0, "MODEL" ], [ 31, 15, 0, 18, 1, "IPADAPTER" ], [ 32, 16, 0, 18, 5, "CLIP_VISION" ] ], "groups": [], "config": {}, "extra": {}, "version": 0.4 }