mirror of
https://github.com/comfyanonymous/ComfyUI.git
synced 2026-01-10 06:10:50 +08:00
142 lines
2.8 KiB
JSON
142 lines
2.8 KiB
JSON
{
|
|
"3": {
|
|
"inputs": {
|
|
"seed": 134898694445843,
|
|
"steps": 1,
|
|
"cfg": 4,
|
|
"sampler_name": "euler",
|
|
"scheduler": "simple",
|
|
"denoise": 1,
|
|
"model": [
|
|
"13",
|
|
0
|
|
],
|
|
"positive": [
|
|
"6",
|
|
0
|
|
],
|
|
"negative": [
|
|
"7",
|
|
0
|
|
],
|
|
"latent_image": [
|
|
"12",
|
|
0
|
|
]
|
|
},
|
|
"class_type": "KSampler",
|
|
"_meta": {
|
|
"title": "KSampler"
|
|
}
|
|
},
|
|
"6": {
|
|
"inputs": {
|
|
"text": "aesthetic 2, anime rpg game style, cute anime girl with gigantic fennec ears and a big fluffy fox tail with long wavy blonde hair and large blue eyes blonde colored eyelashes wearing a pink sweater a large oversized gold trimmed black winter coat and a long blue maxi skirt and large winter boots and a red scarf and large gloves dirty clothes muddy clothes, she is happy and holding a sword in a winter forest with evergreen trees there are the beautiful snow mountains in the background",
|
|
"clip": [
|
|
"11",
|
|
0
|
|
]
|
|
},
|
|
"class_type": "CLIPTextEncode",
|
|
"_meta": {
|
|
"title": "CLIP Text Encode (Prompt)"
|
|
}
|
|
},
|
|
"7": {
|
|
"inputs": {
|
|
"text": "low quality, bad anatomy, extra digits, missing digits, extra limbs, missing limbs",
|
|
"clip": [
|
|
"11",
|
|
0
|
|
]
|
|
},
|
|
"class_type": "CLIPTextEncode",
|
|
"_meta": {
|
|
"title": "CLIP Text Encode (Prompt)"
|
|
}
|
|
},
|
|
"8": {
|
|
"inputs": {
|
|
"samples": [
|
|
"3",
|
|
0
|
|
],
|
|
"vae": [
|
|
"15",
|
|
0
|
|
]
|
|
},
|
|
"class_type": "VAEDecode",
|
|
"_meta": {
|
|
"title": "VAE Decode"
|
|
}
|
|
},
|
|
"9": {
|
|
"inputs": {
|
|
"filename_prefix": "ComfyUI",
|
|
"images": [
|
|
"8",
|
|
0
|
|
]
|
|
},
|
|
"class_type": "SaveImage",
|
|
"_meta": {
|
|
"title": "Save Image"
|
|
}
|
|
},
|
|
"10": {
|
|
"inputs": {
|
|
"clip_name": "t5xxl_fp16.safetensors",
|
|
"type": "chroma",
|
|
"device": "default"
|
|
},
|
|
"class_type": "CLIPLoader",
|
|
"_meta": {
|
|
"title": "Load CLIP"
|
|
}
|
|
},
|
|
"11": {
|
|
"inputs": {
|
|
"min_padding": 0,
|
|
"min_length": 3,
|
|
"clip": [
|
|
"10",
|
|
0
|
|
]
|
|
},
|
|
"class_type": "T5TokenizerOptions",
|
|
"_meta": {
|
|
"title": "T5TokenizerOptions"
|
|
}
|
|
},
|
|
"12": {
|
|
"inputs": {
|
|
"width": 1024,
|
|
"height": 1024,
|
|
"batch_size": 1
|
|
},
|
|
"class_type": "EmptySD3LatentImage",
|
|
"_meta": {
|
|
"title": "EmptySD3LatentImage"
|
|
}
|
|
},
|
|
"13": {
|
|
"inputs": {
|
|
"unet_name": "chroma-unlocked-v37.safetensors",
|
|
"weight_dtype": "default"
|
|
},
|
|
"class_type": "UNETLoader",
|
|
"_meta": {
|
|
"title": "Load Diffusion Model"
|
|
}
|
|
},
|
|
"15": {
|
|
"inputs": {
|
|
"vae_name": "ae.safetensors"
|
|
},
|
|
"class_type": "VAELoader",
|
|
"_meta": {
|
|
"title": "Load VAE"
|
|
}
|
|
}
|
|
} |