This commit is contained in:
lin-bot23 2026-05-08 23:31:01 +08:00 committed by GitHub
commit 2c94271c85
No known key found for this signature in database
GPG Key ID: B5690EEEBB952194
51 changed files with 157 additions and 99 deletions

View File

@ -431,7 +431,8 @@
"extra": { "extra": {
"workflowRendererVersion": "LG" "workflowRendererVersion": "LG"
}, },
"category": "Image Tools/Color adjust" "category": "Image Tools/Color adjust",
"description": "Adjusts image brightness and contrast using a real-time GPU fragment shader."
} }
] ]
}, },

View File

@ -162,7 +162,7 @@
}, },
"revision": 0, "revision": 0,
"config": {}, "config": {},
"name": "local-Canny to Image (Z-Image-Turbo)", "name": "Canny to Image (Z-Image-Turbo)",
"inputNode": { "inputNode": {
"id": -10, "id": -10,
"bounding": [ "bounding": [
@ -1553,7 +1553,8 @@
"VHS_MetadataImage": true, "VHS_MetadataImage": true,
"VHS_KeepIntermediate": true "VHS_KeepIntermediate": true
}, },
"category": "Image generation and editing/Canny to image" "category": "Image generation and editing/Canny to image",
"description": "Generates an image from a Canny edge map using Z-Image-Turbo, with text conditioning."
} }
] ]
}, },

View File

@ -192,7 +192,7 @@
}, },
"revision": 0, "revision": 0,
"config": {}, "config": {},
"name": "local-Canny to Video (LTX 2.0)", "name": "Canny to Video (LTX 2.0)",
"inputNode": { "inputNode": {
"id": -10, "id": -10,
"bounding": [ "bounding": [
@ -3600,7 +3600,8 @@
"extra": { "extra": {
"workflowRendererVersion": "LG" "workflowRendererVersion": "LG"
}, },
"category": "Video generation and editing/Canny to video" "category": "Video generation and editing/Canny to video",
"description": "Generates video from Canny edge maps using LTX-2, with optional synchronized audio."
} }
] ]
}, },

View File

@ -377,7 +377,8 @@
"extra": { "extra": {
"workflowRendererVersion": "LG" "workflowRendererVersion": "LG"
}, },
"category": "Image Tools/Color adjust" "category": "Image Tools/Color adjust",
"description": "Adds lens-style chromatic aberration (color fringing) using a real-time GPU fragment shader."
} }
] ]
} }

View File

@ -256,7 +256,7 @@
"Node name for S&R": "GLSLShader" "Node name for S&R": "GLSLShader"
}, },
"widgets_values": [ "widgets_values": [
"#version 300 es\nprecision highp float;\n\nuniform sampler2D u_image0;\nuniform float u_float0; // temperature (-100 to 100)\nuniform float u_float1; // tint (-100 to 100)\nuniform float u_float2; // vibrance (-100 to 100)\nuniform float u_float3; // saturation (-100 to 100)\n\nin vec2 v_texCoord;\nout vec4 fragColor;\n\nconst float INPUT_SCALE = 0.01;\nconst float TEMP_TINT_PRIMARY = 0.3;\nconst float TEMP_TINT_SECONDARY = 0.15;\nconst float VIBRANCE_BOOST = 2.0;\nconst float SATURATION_BOOST = 2.0;\nconst float SKIN_PROTECTION = 0.5;\nconst float EPSILON = 0.001;\nconst vec3 LUMA_WEIGHTS = vec3(0.299, 0.587, 0.114);\n\nvoid main() {\n vec4 tex = texture(u_image0, v_texCoord);\n vec3 color = tex.rgb;\n \n // Scale inputs: -100/100 \u2192 -1/1\n float temperature = u_float0 * INPUT_SCALE;\n float tint = u_float1 * INPUT_SCALE;\n float vibrance = u_float2 * INPUT_SCALE;\n float saturation = u_float3 * INPUT_SCALE;\n \n // Temperature (warm/cool): positive = warm, negative = cool\n color.r += temperature * TEMP_TINT_PRIMARY;\n color.b -= temperature * TEMP_TINT_PRIMARY;\n \n // Tint (green/magenta): positive = green, negative = magenta\n color.g += tint * TEMP_TINT_PRIMARY;\n color.r -= tint * TEMP_TINT_SECONDARY;\n color.b -= tint * TEMP_TINT_SECONDARY;\n \n // Single clamp after temperature/tint\n color = clamp(color, 0.0, 1.0);\n \n // Vibrance with skin protection\n if (vibrance != 0.0) {\n float maxC = max(color.r, max(color.g, color.b));\n float minC = min(color.r, min(color.g, color.b));\n float sat = maxC - minC;\n float gray = dot(color, LUMA_WEIGHTS);\n \n if (vibrance < 0.0) {\n // Desaturate: -100 \u2192 gray\n color = mix(vec3(gray), color, 1.0 + vibrance);\n } else {\n // Boost less saturated colors more\n float vibranceAmt = vibrance * (1.0 - sat);\n \n // Branchless skin tone protection\n float isWarmTone = step(color.b, color.g) * step(color.g, color.r);\n float warmth = (color.r - color.b) / max(maxC, EPSILON);\n float skinTone = isWarmTone * warmth * sat * (1.0 - sat);\n vibranceAmt *= (1.0 - skinTone * SKIN_PROTECTION);\n \n color = mix(vec3(gray), color, 1.0 + vibranceAmt * VIBRANCE_BOOST);\n }\n }\n \n // Saturation\n if (saturation != 0.0) {\n float gray = dot(color, LUMA_WEIGHTS);\n float satMix = saturation < 0.0\n ? 1.0 + saturation // -100 \u2192 gray\n : 1.0 + saturation * SATURATION_BOOST; // +100 \u2192 3x boost\n color = mix(vec3(gray), color, satMix);\n }\n \n fragColor = vec4(clamp(color, 0.0, 1.0), tex.a);\n}", "#version 300 es\nprecision highp float;\n\nuniform sampler2D u_image0;\nuniform float u_float0; // temperature (-100 to 100)\nuniform float u_float1; // tint (-100 to 100)\nuniform float u_float2; // vibrance (-100 to 100)\nuniform float u_float3; // saturation (-100 to 100)\n\nin vec2 v_texCoord;\nout vec4 fragColor;\n\nconst float INPUT_SCALE = 0.01;\nconst float TEMP_TINT_PRIMARY = 0.3;\nconst float TEMP_TINT_SECONDARY = 0.15;\nconst float VIBRANCE_BOOST = 2.0;\nconst float SATURATION_BOOST = 2.0;\nconst float SKIN_PROTECTION = 0.5;\nconst float EPSILON = 0.001;\nconst vec3 LUMA_WEIGHTS = vec3(0.299, 0.587, 0.114);\n\nvoid main() {\n vec4 tex = texture(u_image0, v_texCoord);\n vec3 color = tex.rgb;\n \n // Scale inputs: -100/100 -1/1\n float temperature = u_float0 * INPUT_SCALE;\n float tint = u_float1 * INPUT_SCALE;\n float vibrance = u_float2 * INPUT_SCALE;\n float saturation = u_float3 * INPUT_SCALE;\n \n // Temperature (warm/cool): positive = warm, negative = cool\n color.r += temperature * TEMP_TINT_PRIMARY;\n color.b -= temperature * TEMP_TINT_PRIMARY;\n \n // Tint (green/magenta): positive = green, negative = magenta\n color.g += tint * TEMP_TINT_PRIMARY;\n color.r -= tint * TEMP_TINT_SECONDARY;\n color.b -= tint * TEMP_TINT_SECONDARY;\n \n // Single clamp after temperature/tint\n color = clamp(color, 0.0, 1.0);\n \n // Vibrance with skin protection\n if (vibrance != 0.0) {\n float maxC = max(color.r, max(color.g, color.b));\n float minC = min(color.r, min(color.g, color.b));\n float sat = maxC - minC;\n float gray = dot(color, LUMA_WEIGHTS);\n \n if (vibrance < 0.0) {\n // Desaturate: -100 gray\n color = mix(vec3(gray), color, 1.0 + vibrance);\n } else {\n // Boost less saturated colors more\n float vibranceAmt = vibrance * (1.0 - sat);\n \n // Branchless skin tone protection\n float isWarmTone = step(color.b, color.g) * step(color.g, color.r);\n float warmth = (color.r - color.b) / max(maxC, EPSILON);\n float skinTone = isWarmTone * warmth * sat * (1.0 - sat);\n vibranceAmt *= (1.0 - skinTone * SKIN_PROTECTION);\n \n color = mix(vec3(gray), color, 1.0 + vibranceAmt * VIBRANCE_BOOST);\n }\n }\n \n // Saturation\n if (saturation != 0.0) {\n float gray = dot(color, LUMA_WEIGHTS);\n float satMix = saturation < 0.0\n ? 1.0 + saturation // -100 → gray\n : 1.0 + saturation * SATURATION_BOOST; // +100 → 3x boost\n color = mix(vec3(gray), color, satMix);\n }\n \n fragColor = vec4(clamp(color, 0.0, 1.0), tex.a);\n}",
"from_input" "from_input"
] ]
}, },
@ -596,7 +596,8 @@
"extra": { "extra": {
"workflowRendererVersion": "LG" "workflowRendererVersion": "LG"
}, },
"category": "Image Tools/Color adjust" "category": "Image Tools/Color adjust",
"description": "Adjusts saturation, temperature, and tint using a real-time GPU fragment shader."
} }
] ]
} }

View File

@ -1129,7 +1129,8 @@
"extra": { "extra": {
"workflowRendererVersion": "LG" "workflowRendererVersion": "LG"
}, },
"category": "Image Tools/Color adjust" "category": "Image Tools/Color adjust",
"description": "Balances colors across shadows, midtones, and highlights using a real-time GPU fragment shader."
} }
] ]
} }

View File

@ -608,7 +608,8 @@
"extra": { "extra": {
"workflowRendererVersion": "LG" "workflowRendererVersion": "LG"
}, },
"category": "Image Tools/Color adjust" "category": "Image Tools/Color adjust",
"description": "Fine-tunes tone and color with per-channel curve adjustments using a real-time GPU fragment shader."
} }
] ]
} }

View File

@ -1609,7 +1609,8 @@
} }
], ],
"extra": {}, "extra": {},
"category": "Image Tools/Crop" "category": "Image Tools/Crop",
"description": "Splits an image into a 2×2 grid of four equal tiles."
} }
] ]
}, },

View File

@ -2946,7 +2946,8 @@
} }
], ],
"extra": {}, "extra": {},
"category": "Image Tools/Crop" "category": "Image Tools/Crop",
"description": "Splits an image into a 3×3 grid of nine equal tiles."
} }
] ]
}, },

View File

@ -1579,7 +1579,8 @@
"VHS_MetadataImage": true, "VHS_MetadataImage": true,
"VHS_KeepIntermediate": true "VHS_KeepIntermediate": true
}, },
"category": "Image generation and editing/Depth to image" "category": "Image generation and editing/Depth to image",
"description": "Generates an image from a depth map using Z-Image-Turbo with text conditioning."
}, },
{ {
"id": "458bdf3c-4b58-421c-af50-c9c663a4d74c", "id": "458bdf3c-4b58-421c-af50-c9c663a4d74c",
@ -2461,7 +2462,8 @@
] ]
}, },
"workflowRendererVersion": "LG" "workflowRendererVersion": "LG"
} },
"description": "Estimates a monocular depth map from an input image using the Lotus depth estimation model."
} }
] ]
}, },

View File

@ -4233,7 +4233,8 @@
"extra": { "extra": {
"workflowRendererVersion": "LG" "workflowRendererVersion": "LG"
}, },
"category": "Video generation and editing/Depth to video" "category": "Video generation and editing/Depth to video",
"description": "Generates video from depth maps using LTX-2, with optional synchronized audio."
}, },
{ {
"id": "38b60539-50a7-42f9-a5fe-bdeca26272e2", "id": "38b60539-50a7-42f9-a5fe-bdeca26272e2",
@ -5192,7 +5193,8 @@
], ],
"extra": { "extra": {
"workflowRendererVersion": "LG" "workflowRendererVersion": "LG"
} },
"description": "Estimates a monocular depth map from an input image using the Lotus depth estimation model."
} }
] ]
}, },

View File

@ -450,7 +450,8 @@
"extra": { "extra": {
"workflowRendererVersion": "LG" "workflowRendererVersion": "LG"
}, },
"category": "Image Tools/Blur" "category": "Image Tools/Blur",
"description": "Applies bilateral (edge-preserving) blur to soften images while retaining detail."
} }
] ]
}, },

View File

@ -580,7 +580,8 @@
"extra": { "extra": {
"workflowRendererVersion": "LG" "workflowRendererVersion": "LG"
}, },
"category": "Image Tools/Color adjust" "category": "Image Tools/Color adjust",
"description": "Adds procedural film grain texture for a cinematic look via GPU fragment shader."
} }
] ]
} }

View File

@ -3350,7 +3350,8 @@
} }
], ],
"extra": {}, "extra": {},
"category": "Video generation and editing/First-Last-Frame to Video" "category": "Video generation and editing/First-Last-Frame to Video",
"description": "Generates a video interpolating between first and last keyframes using LTX-2.3."
} }
] ]
}, },

View File

@ -575,7 +575,8 @@
"extra": { "extra": {
"workflowRendererVersion": "LG" "workflowRendererVersion": "LG"
}, },
"category": "Image Tools/Color adjust" "category": "Image Tools/Color adjust",
"description": "Adds a glow/bloom effect around bright image areas via GPU fragment shader."
} }
] ]
} }

View File

@ -752,7 +752,8 @@
"extra": { "extra": {
"workflowRendererVersion": "LG" "workflowRendererVersion": "LG"
}, },
"category": "Image Tools/Color adjust" "category": "Image Tools/Color adjust",
"description": "Adjusts hue, saturation, and lightness of an image using a real-time GPU fragment shader."
} }
] ]
} }

View File

@ -374,7 +374,8 @@
"extra": { "extra": {
"workflowRendererVersion": "LG" "workflowRendererVersion": "LG"
}, },
"category": "Image Tools/Blur" "category": "Image Tools/Blur",
"description": "Applies Gaussian blur to soften an image or simulate depth-of-field effects."
} }
] ]
} }

View File

@ -310,7 +310,8 @@
"extra": { "extra": {
"workflowRendererVersion": "LG" "workflowRendererVersion": "LG"
}, },
"category": "Text generation/Image Captioning" "category": "Text generation/Image Captioning",
"description": "Generates descriptive captions for images using Google Gemini's multimodal LLM."
} }
] ]
} }

View File

@ -315,7 +315,8 @@
"extra": { "extra": {
"workflowRendererVersion": "LG" "workflowRendererVersion": "LG"
}, },
"category": "Image Tools/Color adjust" "category": "Image Tools/Color adjust",
"description": "Manipulates individual RGBA channels for masking, compositing, and channel effects."
} }
] ]
} }

View File

@ -2138,7 +2138,8 @@
"extra": { "extra": {
"workflowRendererVersion": "LG" "workflowRendererVersion": "LG"
}, },
"category": "Image generation and editing/Edit image" "category": "Image generation and editing/Edit image",
"description": "Edits images via text instructions using FireRed Image Edit 1.1, a diffusion-based instruction-following editing model."
} }
] ]
}, },

View File

@ -1472,7 +1472,8 @@
"extra": { "extra": {
"workflowRendererVersion": "LG" "workflowRendererVersion": "LG"
}, },
"category": "Image generation and editing/Edit image" "category": "Image generation and editing/Edit image",
"description": "Edits images via text instructions using FLUX.2 [klein] 4B, supporting both T2I and image editing."
}, },
{ {
"id": "6007e698-2ebd-4917-84d8-299b35d7b7ab", "id": "6007e698-2ebd-4917-84d8-299b35d7b7ab",
@ -1821,7 +1822,8 @@
], ],
"extra": { "extra": {
"workflowRendererVersion": "LG" "workflowRendererVersion": "LG"
} },
"description": "Applies reference image conditioning for style/identity transfer (Flux.2 Klein 4B)."
} }
] ]
}, },

View File

@ -1417,7 +1417,8 @@
} }
], ],
"extra": {}, "extra": {},
"category": "Image generation and editing/Edit image" "category": "Image generation and editing/Edit image",
"description": "Edits images via text instructions using LongCat Image Edit, an instruction-following image editing diffusion model."
} }
] ]
}, },

View File

@ -132,7 +132,7 @@
}, },
"revision": 0, "revision": 0,
"config": {}, "config": {},
"name": "local-Image Edit (Qwen 2511)", "name": "Image Edit (Qwen 2511)",
"inputNode": { "inputNode": {
"id": -10, "id": -10,
"bounding": [ "bounding": [
@ -1468,7 +1468,8 @@
"VHS_MetadataImage": true, "VHS_MetadataImage": true,
"VHS_KeepIntermediate": true "VHS_KeepIntermediate": true
}, },
"category": "Image generation and editing/Edit image" "category": "Image generation and editing/Edit image",
"description": "Edits images via text instructions using Qwen-Image-Edit-2511 with improved character consistency and integrated LoRA."
} }
] ]
}, },

View File

@ -1188,7 +1188,8 @@
"extra": { "extra": {
"workflowRendererVersion": "LG" "workflowRendererVersion": "LG"
}, },
"category": "Image generation and editing/Inpaint image" "category": "Image generation and editing/Inpaint image",
"description": "Inpaints masked image regions using Flux.1 fill [dev], BFL's inpainting/outpainting model."
} }
] ]
}, },

View File

@ -1548,7 +1548,8 @@
"extra": { "extra": {
"workflowRendererVersion": "LG" "workflowRendererVersion": "LG"
}, },
"category": "Image generation and editing/Inpaint image" "category": "Image generation and editing/Inpaint image",
"description": "Inpaints masked regions using Qwen-Image, extending its multilingual text rendering to inpainting tasks."
}, },
{ {
"id": "56a1f603-fbd2-40ed-94ef-c9ecbd96aca8", "id": "56a1f603-fbd2-40ed-94ef-c9ecbd96aca8",
@ -1907,7 +1908,8 @@
], ],
"extra": { "extra": {
"workflowRendererVersion": "LG" "workflowRendererVersion": "LG"
} },
"description": "Expands and softens mask edges to reduce visible seams after image processing."
} }
] ]
}, },

View File

@ -742,7 +742,8 @@
"extra": { "extra": {
"workflowRendererVersion": "LG" "workflowRendererVersion": "LG"
}, },
"category": "Image Tools/Color adjust" "category": "Image Tools/Color adjust",
"description": "Adjusts black point, white point, and gamma for tonal range control via GPU shader."
} }
] ]
}, },

View File

@ -1919,7 +1919,8 @@
"extra": { "extra": {
"workflowRendererVersion": "LG" "workflowRendererVersion": "LG"
}, },
"category": "Image generation and editing/Outpaint image" "category": "Image generation and editing/Outpaint image",
"description": "Outpaints beyond image boundaries using Qwen-Image's outpainting capabilities."
}, },
{ {
"id": "f93c215e-c393-460e-9534-ed2c3d8a652e", "id": "f93c215e-c393-460e-9534-ed2c3d8a652e",
@ -2278,7 +2279,8 @@
], ],
"extra": { "extra": {
"workflowRendererVersion": "LG" "workflowRendererVersion": "LG"
} },
"description": "Expands and softens mask edges to reduce visible seams after image processing."
}, },
{ {
"id": "2a4b2cc0-db37-4302-a067-da392f38f06b", "id": "2a4b2cc0-db37-4302-a067-da392f38f06b",
@ -2733,7 +2735,8 @@
], ],
"extra": { "extra": {
"workflowRendererVersion": "LG" "workflowRendererVersion": "LG"
} },
"description": "Scales both image and mask together while preserving alignment for editing workflows."
} }
] ]
}, },

View File

@ -141,7 +141,7 @@
}, },
"revision": 0, "revision": 0,
"config": {}, "config": {},
"name": "local-Image Upscale(Z-image-Turbo)", "name": "Image Upscale(Z-image-Turbo)",
"inputNode": { "inputNode": {
"id": -10, "id": -10,
"bounding": [ "bounding": [
@ -1302,7 +1302,8 @@
"extra": { "extra": {
"workflowRendererVersion": "LG" "workflowRendererVersion": "LG"
}, },
"category": "Image generation and editing/Enhance" "category": "Image generation and editing/Enhance",
"description": "Upscales images to higher resolution using Z-Image-Turbo."
} }
] ]
}, },

View File

@ -99,7 +99,7 @@
}, },
"revision": 0, "revision": 0,
"config": {}, "config": {},
"name": "local-Image to Depth Map (Lotus)", "name": "Image to Depth Map (Lotus)",
"inputNode": { "inputNode": {
"id": -10, "id": -10,
"bounding": [ "bounding": [
@ -948,7 +948,8 @@
"extra": { "extra": {
"workflowRendererVersion": "LG" "workflowRendererVersion": "LG"
}, },
"category": "Image generation and editing/Depth to image" "category": "Image generation and editing/Depth to image",
"description": "Estimates a monocular depth map from an input image using the Lotus depth estimation model."
} }
] ]
}, },

View File

@ -1586,7 +1586,8 @@
"extra": { "extra": {
"workflowRendererVersion": "LG" "workflowRendererVersion": "LG"
}, },
"category": "Image generation and editing/Image to layers" "category": "Image generation and editing/Image to layers",
"description": "Decomposes an image into variable-resolution RGBA layers for independent editing using Qwen-Image-Layered."
} }
] ]
}, },

View File

@ -72,7 +72,7 @@
}, },
"revision": 0, "revision": 0,
"config": {}, "config": {},
"name": "local-Image to Model (Hunyuan3d 2.1)", "name": "Image to Model (Hunyuan3d 2.1)",
"inputNode": { "inputNode": {
"id": -10, "id": -10,
"bounding": [ "bounding": [
@ -765,7 +765,8 @@
"extra": { "extra": {
"workflowRendererVersion": "LG" "workflowRendererVersion": "LG"
}, },
"category": "3D/Image to 3D Model" "category": "3D/Image to 3D Model",
"description": "Generates 3D mesh models from a single input image using Hunyuan3D 2.0/2.1."
} }
] ]
}, },

View File

@ -4223,7 +4223,8 @@
"extra": { "extra": {
"workflowRendererVersion": "Vue-corrected" "workflowRendererVersion": "Vue-corrected"
}, },
"category": "Video generation and editing/Image to video" "category": "Video generation and editing/Image to video",
"description": "Generates video from a single input image using LTX-2.3."
} }
] ]
}, },

View File

@ -206,7 +206,7 @@
}, },
"revision": 0, "revision": 0,
"config": {}, "config": {},
"name": "local-Image to Video (Wan 2.2)", "name": "Image to Video (Wan 2.2)",
"inputNode": { "inputNode": {
"id": -10, "id": -10,
"bounding": [ "bounding": [
@ -2027,7 +2027,8 @@
"extra": { "extra": {
"workflowRendererVersion": "LG" "workflowRendererVersion": "LG"
}, },
"category": "Video generation and editing/Image to video" "category": "Video generation and editing/Image to video",
"description": "Generates video from an image and text prompt using Wan2.2, supporting T2V and I2V."
} }
] ]
}, },

View File

@ -134,7 +134,7 @@
}, },
"revision": 0, "revision": 0,
"config": {}, "config": {},
"name": "local-Pose to Image (Z-Image-Turbo)", "name": "Pose to Image (Z-Image-Turbo)",
"inputNode": { "inputNode": {
"id": -10, "id": -10,
"bounding": [ "bounding": [
@ -1298,7 +1298,8 @@
"VHS_MetadataImage": true, "VHS_MetadataImage": true,
"VHS_KeepIntermediate": true "VHS_KeepIntermediate": true
}, },
"category": "Image generation and editing/Pose to image" "category": "Image generation and editing/Pose to image",
"description": "Generates an image from pose keypoints using Z-Image-Turbo with text conditioning."
} }
] ]
}, },

View File

@ -3870,7 +3870,8 @@
"extra": { "extra": {
"workflowRendererVersion": "LG" "workflowRendererVersion": "LG"
}, },
"category": "Video generation and editing/Pose to video" "category": "Video generation and editing/Pose to video",
"description": "Generates video from pose reference frames using LTX-2, with optional synchronized audio."
} }
] ]
}, },

View File

@ -270,7 +270,8 @@
"extra": { "extra": {
"workflowRendererVersion": "LG" "workflowRendererVersion": "LG"
}, },
"category": "Text generation/Prompt enhance" "category": "Text generation/Prompt enhance",
"description": "Expands short text prompts into detailed descriptions using a text generation model for better generation quality."
} }
] ]
}, },

View File

@ -302,7 +302,8 @@
"extra": { "extra": {
"workflowRendererVersion": "LG" "workflowRendererVersion": "LG"
}, },
"category": "Image Tools/Sharpen" "category": "Image Tools/Sharpen",
"description": "Sharpens image details using a GPU fragment shader for enhanced clarity."
} }
] ]
} }

View File

@ -222,7 +222,7 @@
}, },
"revision": 0, "revision": 0,
"config": {}, "config": {},
"name": "local-Text to Audio (ACE-Step 1.5)", "name": "Text to Audio (ACE-Step 1.5)",
"inputNode": { "inputNode": {
"id": -10, "id": -10,
"bounding": [ "bounding": [
@ -1502,7 +1502,8 @@
"extra": { "extra": {
"workflowRendererVersion": "LG" "workflowRendererVersion": "LG"
}, },
"category": "Audio/Music generation" "category": "Audio/Music generation",
"description": "Generates audio/music from text prompts using ACE-Step 1.5, a diffusion-based audio generation model."
} }
] ]
}, },

View File

@ -1029,7 +1029,8 @@
"extra": { "extra": {
"workflowRendererVersion": "LG" "workflowRendererVersion": "LG"
}, },
"category": "Image generation and editing/Text to image" "category": "Image generation and editing/Text to image",
"description": "Generates images from text prompts using Flux.1 [dev], BFL's 12B diffusion model."
} }
] ]
}, },

View File

@ -1023,7 +1023,8 @@
"extra": { "extra": {
"workflowRendererVersion": "LG" "workflowRendererVersion": "LG"
}, },
"category": "Image generation and editing/Text to image" "category": "Image generation and editing/Text to image",
"description": "Generates images from text prompts using Flux.1 Krea Dev, a BFL × Krea collaboration variant."
} }
] ]
}, },

View File

@ -1104,7 +1104,8 @@
"extra": { "extra": {
"workflowRendererVersion": "LG" "workflowRendererVersion": "LG"
}, },
"category": "Image generation and editing/Text to image" "category": "Image generation and editing/Text to image",
"description": "Generates images from text prompts using NetaYume Lumina, a Lumina-Next variant fine-tuned for anime-style and illustration generation."
}, },
{ {
"id": "a07fdf06-1bda-4dac-bdbd-63ee8ebca1c9", "id": "a07fdf06-1bda-4dac-bdbd-63ee8ebca1c9",
@ -1458,7 +1459,8 @@
], ],
"extra": { "extra": {
"workflowRendererVersion": "LG" "workflowRendererVersion": "LG"
} },
"description": "Encodes a negative text prompt via CLIP for classifier-free guidance in anime-style generation (NetaYume Lumina)."
} }
] ]
}, },

View File

@ -1941,7 +1941,8 @@
"extra": { "extra": {
"workflowRendererVersion": "Vue-corrected" "workflowRendererVersion": "Vue-corrected"
}, },
"category": "Image generation and editing/Text to image" "category": "Image generation and editing/Text to image",
"description": "Generates images from text prompts using Qwen-Image-2512, with enhanced human realism and finer natural detail over the base version."
} }
] ]
}, },

View File

@ -1873,7 +1873,8 @@
"extra": { "extra": {
"workflowRendererVersion": "LG" "workflowRendererVersion": "LG"
}, },
"category": "Image generation and editing/Text to image" "category": "Image generation and editing/Text to image",
"description": "Generates images from text prompts using Qwen-Image, Alibaba's 20B MMDiT model with excellent multilingual text rendering."
} }
] ]
}, },

View File

@ -149,7 +149,7 @@
}, },
"revision": 0, "revision": 0,
"config": {}, "config": {},
"name": "local-Text to Image (Z-Image-Turbo)", "name": "Text to Image (Z-Image-Turbo)",
"inputNode": { "inputNode": {
"id": -10, "id": -10,
"bounding": [ "bounding": [
@ -1054,7 +1054,8 @@
"extra": { "extra": {
"workflowRendererVersion": "LG" "workflowRendererVersion": "LG"
}, },
"category": "Image generation and editing/Text to image" "category": "Image generation and editing/Text to image",
"description": "Generates images from text prompts using Z-Image-Turbo, Alibaba's distilled 6B DiT model."
} }
] ]
}, },

View File

@ -4286,7 +4286,8 @@
"extra": { "extra": {
"workflowRendererVersion": "Vue-corrected" "workflowRendererVersion": "Vue-corrected"
}, },
"category": "Video generation and editing/Text to video" "category": "Video generation and editing/Text to video",
"description": "Generates video from text prompts using LTX-2.3, Lightricks' video diffusion model."
} }
] ]
}, },

View File

@ -1572,7 +1572,8 @@
"extra": { "extra": {
"workflowRendererVersion": "LG" "workflowRendererVersion": "LG"
}, },
"category": "Video generation and editing/Text to video" "category": "Video generation and editing/Text to video",
"description": "Generates video from text prompts using Wan2.2, Alibaba's diffusion video model."
} }
] ]
}, },

View File

@ -434,7 +434,8 @@
"extra": { "extra": {
"workflowRendererVersion": "LG" "workflowRendererVersion": "LG"
}, },
"category": "Image Tools/Sharpen" "category": "Image Tools/Sharpen",
"description": "Enhances edge contrast via unsharp masking for a sharper image appearance."
} }
] ]
} }

View File

@ -307,7 +307,8 @@
"extra": { "extra": {
"workflowRendererVersion": "LG" "workflowRendererVersion": "LG"
}, },
"category": "Text generation/Video Captioning" "category": "Text generation/Video Captioning",
"description": "Generates descriptive captions for video input using Google Gemini's multimodal LLM."
} }
] ]
} }

View File

@ -165,7 +165,7 @@
}, },
"revision": 0, "revision": 0,
"config": {}, "config": {},
"name": "local-Video Inpaint(Wan2.1 VACE)", "name": "Video Inpaint(Wan2.1 VACE)",
"inputNode": { "inputNode": {
"id": -10, "id": -10,
"bounding": [ "bounding": [
@ -2368,7 +2368,8 @@
"extra": { "extra": {
"workflowRendererVersion": "LG" "workflowRendererVersion": "LG"
}, },
"category": "Video generation and editing/Inpaint video" "category": "Video generation and editing/Inpaint video",
"description": "Inpaints masked regions in video frames using Wan2.1 VACE."
} }
] ]
}, },

View File

@ -584,7 +584,8 @@
"extra": { "extra": {
"workflowRendererVersion": "LG" "workflowRendererVersion": "LG"
}, },
"category": "Video Tools/Stitch videos" "category": "Video Tools/Stitch videos",
"description": "Stitches multiple video clips into a single sequential video file."
} }
] ]
} }

View File

@ -412,7 +412,8 @@
"extra": { "extra": {
"workflowRendererVersion": "LG" "workflowRendererVersion": "LG"
}, },
"category": "Video generation and editing/Enhance video" "category": "Video generation and editing/Enhance video",
"description": "Upscales video to 4× resolution using a GAN-based upscaling model."
} }
] ]
}, },