Move palette to separate node

Add default value: black and white
Ensure that RGB values are <= 255
Change palette input to multi-line. Colors can be divided per line, in parentheses or not.
This commit is contained in:
missionfloyd 2023-04-08 03:22:08 -06:00 committed by GitHub
parent 809a1dd23d
commit 0ceb7f3009
No known key found for this signature in database
GPG Key ID: 4AEE18F83AFDEB23

View File

@ -3,6 +3,7 @@ import torch
import torch.nn.functional as F
from PIL import Image, ImageColor
import re
import itertools
import comfy.utils
@ -125,7 +126,6 @@ class Quantize:
"max": 256,
"step": 1
}),
"palette": ("STRING", {"default": ""}),
"dither": (["none", "floyd-steinberg"],),
},
}
@ -135,7 +135,7 @@ class Quantize:
CATEGORY = "image/postprocessing"
def quantize(self, palette: str, image: torch.Tensor, colors: int = 256, dither: str = "FLOYDSTEINBERG"):
def quantize(self, image: torch.Tensor, colors: int = 256, dither: str = "FLOYDSTEINBERG"):
batch_size, height, width, _ = image.shape
result = torch.zeros_like(image)
@ -146,21 +146,7 @@ class Quantize:
img = (tensor_image * 255).to(torch.uint8).numpy()
pil_image = Image.fromarray(img, mode='RGB')
if palette:
def flatten_list(list_of_lists, flat_list=[]):
for item in list_of_lists:
if type(item) == tuple:
flatten_list(item, flat_list)
else:
flat_list.append(item)
return flat_list
pal_img = Image.new('P', (1, 1))
pal_colors = palette.replace(" ", "").split(",")
pal_colors = map(lambda i: ImageColor.getrgb(i) if re.search("#[a-fA-F0-9]{6}", i) else int(i), pal_colors)
pal_img.putpalette(flatten_list(pal_colors))
else:
pal_img = pil_image.quantize(colors=colors) # Required as described in https://github.com/python-pillow/Pillow/issues/5836
pal_img = pil_image.quantize(colors=colors) # Required as described in https://github.com/python-pillow/Pillow/issues/5836
quantized_image = pil_image.quantize(colors=colors, palette=pal_img, dither=dither_option)
quantized_array = torch.tensor(np.array(quantized_image.convert("RGB"))).float() / 255
@ -168,6 +154,60 @@ class Quantize:
return (result,)
class QuantizePalette:
def __init__(self):
pass
@classmethod
def INPUT_TYPES(s):
return {
"required": {
"image": ("IMAGE",),
"palette": ("STRING", {"default": "#000000, #ffffff", "multiline": True}),
"dither": (["none", "floyd-steinberg"],),
},
}
RETURN_TYPES = ("IMAGE",)
FUNCTION = "quantize"
CATEGORY = "image/postprocessing"
def quantize(self, palette: str, image: torch.Tensor, dither: str = "FLOYDSTEINBERG"):
batch_size, height, width, _ = image.shape
result = torch.zeros_like(image)
for b in range(batch_size):
tensor_image = image[b]
img = (tensor_image * 255).to(torch.uint8).numpy()
pil_image = Image.fromarray(img, mode='RGB')
if not palette:
raise ValueError("Palette is empty")
def parse_palette(color_str):
if re.match(r'#[a-fA-F0-9]{6}', color_str):
return ImageColor.getrgb(color_str)
color_rgb = re.match(r'\(?(\d{1,3}),(\d{1,3}),(\d{1,3})\)?', color_str)
if color_rgb and int(color_rgb.group(1)) <= 255 and int(color_rgb.group(2)) <= 255 and int(color_rgb.group(3)) <= 255:
return tuple(map(int, re.findall(r'\d{1,3}', color_str)))
else:
raise ValueError(f"Invalid color format: {color_str}")
pal_img = Image.new('P', (1, 1))
pal_colors = palette.replace(" ", "")
pal_colors = re.findall(r'#[a-fA-F0-9]{6}|\(?\d{1,3},\d{1,3},\d{1,3}\)?', pal_colors)
pal_colors = list(itertools.chain.from_iterable(map(parse_palette, pal_colors)))
pal_img.putpalette(pal_colors)
quantized_image = pil_image.quantize(palette=pal_img)
quantized_array = torch.tensor(np.array(quantized_image.convert("RGB"))).float() / 255
result[b] = quantized_array
return (result,)
class Sharpen:
def __init__(self):
pass
@ -222,5 +262,6 @@ NODE_CLASS_MAPPINGS = {
"ImageBlend": Blend,
"ImageBlur": Blur,
"ImageQuantize": Quantize,
"ImageQuantizePalette": QuantizePalette,
"ImageSharpen": Sharpen,
}