{ "last_node_id": 15, "last_link_id": 19, "nodes": [ { "id": 7, "type": "CLIPTextEncode", "pos": [413, 389], "size": [425.28, 180.61], "flags": {}, "order": 5, "mode": 0, "inputs": [ { "name": "clip", "type": "CLIP", "link": 18 } ], "outputs": [ { "name": "CONDITIONING", "type": "CONDITIONING", "links": [6], "slot_index": 0 } ], "properties": { "Node name for S&R": "CLIPTextEncode" }, "widgets_values": ["bad hands"] }, { "id": 6, "type": "CLIPTextEncode", "pos": [415, 186], "size": [422.85, 164.31], "flags": {}, "order": 4, "mode": 0, "inputs": [ { "name": "clip", "type": "CLIP", "link": 17 } ], "outputs": [ { "name": "CONDITIONING", "type": "CONDITIONING", "links": [4], "slot_index": 0 } ], "properties": { "Node name for S&R": "CLIPTextEncode" }, "widgets_values": [ "masterpiece best quality girl standing in victorian clothing" ] }, { "id": 8, "type": "VAEDecode", "pos": [1209, 188], "size": [210, 46], "flags": {}, "order": 7, "mode": 0, "inputs": [ { "name": "samples", "type": "LATENT", "link": 7 }, { "name": "vae", "type": "VAE", "link": 19 } ], "outputs": [ { "name": "IMAGE", "type": "IMAGE", "links": [15], "slot_index": 0 } ], "properties": { "Node name for S&R": "VAEDecode" }, "widgets_values": [] }, { "id": 9, "type": "SaveImage", "pos": [1791, 169], "size": [455.99, 553.09], "flags": {}, "order": 9, "mode": 0, "inputs": [ { "name": "images", "type": "IMAGE", "link": 16 } ], "outputs": [], "properties": {}, "widgets_values": ["ComfyUI"] }, { "id": 14, "type": "ImageUpscaleWithModel", "pos": [1506, 151], "size": [241.8, 46], "flags": {}, "order": 8, "mode": 0, "inputs": [ { "name": "upscale_model", "type": "UPSCALE_MODEL", "link": 14 }, { "name": "image", "type": "IMAGE", "link": 15 } ], "outputs": [ { "name": "IMAGE", "type": "IMAGE", "links": [16], "slot_index": 0 } ], "properties": { "Node name for S&R": "ImageUpscaleWithModel" }, "widgets_values": [] }, { "id": 13, "type": "UpscaleModelLoader", "pos": [1128, 51], "size": [315, 58], "flags": {}, "order": 0, "mode": 0, "inputs": [], "outputs": [ { "name": "UPSCALE_MODEL", "type": "UPSCALE_MODEL", "links": [14], "slot_index": 0 } ], "properties": { "Node name for S&R": "UpscaleModelLoader" }, "widgets_values": ["RealESRGAN_x4plus.pth"] }, { "id": 3, "type": "KSampler", "pos": [863, 186], "size": [315, 262], "flags": {}, "order": 6, "mode": 0, "inputs": [ { "name": "model", "type": "MODEL", "link": 1 }, { "name": "positive", "type": "CONDITIONING", "link": 4 }, { "name": "negative", "type": "CONDITIONING", "link": 6 }, { "name": "latent_image", "type": "LATENT", "link": 2 } ], "outputs": [ { "name": "LATENT", "type": "LATENT", "links": [7], "slot_index": 0 } ], "properties": { "Node name for S&R": "KSampler" }, "widgets_values": [ 833543590226030, "randomize", 20, 8, "euler", "normal", 1 ] }, { "id": 4, "type": "CheckpointLoaderSimple", "pos": [-11, 307], "size": [315, 98], "flags": {}, "order": 1, "mode": 0, "inputs": [], "outputs": [ { "name": "MODEL", "type": "MODEL", "links": [1], "slot_index": 0 }, { "name": "CLIP", "type": "CLIP", "links": [17, 18], "slot_index": 1 }, { "name": "VAE", "type": "VAE", "links": [19], "slot_index": 2 } ], "properties": { "Node name for S&R": "CheckpointLoaderSimple" }, "widgets_values": ["v1-5-pruned-emaonly-fp16.safetensors"] }, { "id": 5, "type": "EmptyLatentImage", "pos": [473, 609], "size": [315, 106], "flags": {}, "order": 2, "mode": 0, "inputs": [], "outputs": [ { "name": "LATENT", "type": "LATENT", "links": [2], "slot_index": 0 } ], "properties": { "Node name for S&R": "EmptyLatentImage" }, "widgets_values": [512, 512, 1] }, { "id": 15, "type": "MarkdownNote", "pos": [0, 465], "size": [225, 60], "flags": {}, "order": 3, "mode": 0, "inputs": [], "outputs": [], "properties": {}, "widgets_values": [ "\ud83d\udec8 [Learn more about this workflow](https://comfyanonymous.github.io/ComfyUI_examples/upscale_models/)" ], "color": "#432", "bgcolor": "#653" } ], "links": [ [1, 4, 0, 3, 0, "MODEL"], [2, 5, 0, 3, 3, "LATENT"], [4, 6, 0, 3, 1, "CONDITIONING"], [6, 7, 0, 3, 2, "CONDITIONING"], [7, 3, 0, 8, 0, "LATENT"], [14, 13, 0, 14, 0, "UPSCALE_MODEL"], [15, 8, 0, 14, 1, "IMAGE"], [16, 14, 0, 9, 0, "IMAGE"], [17, 4, 1, 6, 0, "CLIP"], [18, 4, 1, 7, 0, "CLIP"], [19, 4, 2, 8, 1, "VAE"] ], "groups": [], "config": {}, "extra": { "ds": { "scale": 0.82, "offset": [400.67, 431.06] } }, "version": 0.4, "models": [ { "name": "v1-5-pruned-emaonly-fp16.safetensors", "url": "https://huggingface.co/Comfy-Org/stable-diffusion-v1-5-archive/resolve/main/v1-5-pruned-emaonly-fp16.safetensors?download=true", "directory": "checkpoints" }, { "name": "RealESRGAN_x4plus.pth", "url": "https://github.com/xinntao/Real-ESRGAN/releases/download/v0.1.0/RealESRGAN_x4plus.pth", "directory": "upscale_models" } ] }