mirror of
https://github.com/comfyanonymous/ComfyUI.git
synced 2025-12-21 12:00:49 +08:00
chore(api-nodes): by default set Watermark generation to False (#11437)
Some checks are pending
Python Linting / Run Ruff (push) Waiting to run
Python Linting / Run Pylint (push) Waiting to run
Full Comfy CI Workflow Runs / test-stable (12.1, , linux, 3.10, [self-hosted Linux], stable) (push) Waiting to run
Full Comfy CI Workflow Runs / test-stable (12.1, , linux, 3.11, [self-hosted Linux], stable) (push) Waiting to run
Full Comfy CI Workflow Runs / test-stable (12.1, , linux, 3.12, [self-hosted Linux], stable) (push) Waiting to run
Full Comfy CI Workflow Runs / test-unix-nightly (12.1, , linux, 3.11, [self-hosted Linux], nightly) (push) Waiting to run
Execution Tests / test (macos-latest) (push) Waiting to run
Execution Tests / test (ubuntu-latest) (push) Waiting to run
Execution Tests / test (windows-latest) (push) Waiting to run
Test server launches without errors / test (push) Waiting to run
Unit Tests / test (macos-latest) (push) Waiting to run
Unit Tests / test (ubuntu-latest) (push) Waiting to run
Unit Tests / test (windows-2022) (push) Waiting to run
Some checks are pending
Python Linting / Run Ruff (push) Waiting to run
Python Linting / Run Pylint (push) Waiting to run
Full Comfy CI Workflow Runs / test-stable (12.1, , linux, 3.10, [self-hosted Linux], stable) (push) Waiting to run
Full Comfy CI Workflow Runs / test-stable (12.1, , linux, 3.11, [self-hosted Linux], stable) (push) Waiting to run
Full Comfy CI Workflow Runs / test-stable (12.1, , linux, 3.12, [self-hosted Linux], stable) (push) Waiting to run
Full Comfy CI Workflow Runs / test-unix-nightly (12.1, , linux, 3.11, [self-hosted Linux], nightly) (push) Waiting to run
Execution Tests / test (macos-latest) (push) Waiting to run
Execution Tests / test (ubuntu-latest) (push) Waiting to run
Execution Tests / test (windows-latest) (push) Waiting to run
Test server launches without errors / test (push) Waiting to run
Unit Tests / test (macos-latest) (push) Waiting to run
Unit Tests / test (ubuntu-latest) (push) Waiting to run
Unit Tests / test (windows-2022) (push) Waiting to run
This commit is contained in:
parent
fb478f679a
commit
0899012ad6
@ -10,7 +10,7 @@ class Text2ImageTaskCreationRequest(BaseModel):
|
|||||||
size: str | None = Field(None)
|
size: str | None = Field(None)
|
||||||
seed: int | None = Field(0, ge=0, le=2147483647)
|
seed: int | None = Field(0, ge=0, le=2147483647)
|
||||||
guidance_scale: float | None = Field(..., ge=1.0, le=10.0)
|
guidance_scale: float | None = Field(..., ge=1.0, le=10.0)
|
||||||
watermark: bool | None = Field(True)
|
watermark: bool | None = Field(False)
|
||||||
|
|
||||||
|
|
||||||
class Image2ImageTaskCreationRequest(BaseModel):
|
class Image2ImageTaskCreationRequest(BaseModel):
|
||||||
@ -21,7 +21,7 @@ class Image2ImageTaskCreationRequest(BaseModel):
|
|||||||
size: str | None = Field("adaptive")
|
size: str | None = Field("adaptive")
|
||||||
seed: int | None = Field(..., ge=0, le=2147483647)
|
seed: int | None = Field(..., ge=0, le=2147483647)
|
||||||
guidance_scale: float | None = Field(..., ge=1.0, le=10.0)
|
guidance_scale: float | None = Field(..., ge=1.0, le=10.0)
|
||||||
watermark: bool | None = Field(True)
|
watermark: bool | None = Field(False)
|
||||||
|
|
||||||
|
|
||||||
class Seedream4Options(BaseModel):
|
class Seedream4Options(BaseModel):
|
||||||
@ -37,7 +37,7 @@ class Seedream4TaskCreationRequest(BaseModel):
|
|||||||
seed: int = Field(..., ge=0, le=2147483647)
|
seed: int = Field(..., ge=0, le=2147483647)
|
||||||
sequential_image_generation: str = Field("disabled")
|
sequential_image_generation: str = Field("disabled")
|
||||||
sequential_image_generation_options: Seedream4Options = Field(Seedream4Options(max_images=15))
|
sequential_image_generation_options: Seedream4Options = Field(Seedream4Options(max_images=15))
|
||||||
watermark: bool = Field(True)
|
watermark: bool = Field(False)
|
||||||
|
|
||||||
|
|
||||||
class ImageTaskCreationResponse(BaseModel):
|
class ImageTaskCreationResponse(BaseModel):
|
||||||
|
|||||||
@ -112,7 +112,7 @@ class ByteDanceImageNode(IO.ComfyNode):
|
|||||||
),
|
),
|
||||||
IO.Boolean.Input(
|
IO.Boolean.Input(
|
||||||
"watermark",
|
"watermark",
|
||||||
default=True,
|
default=False,
|
||||||
tooltip='Whether to add an "AI generated" watermark to the image',
|
tooltip='Whether to add an "AI generated" watermark to the image',
|
||||||
optional=True,
|
optional=True,
|
||||||
),
|
),
|
||||||
@ -215,7 +215,7 @@ class ByteDanceImageEditNode(IO.ComfyNode):
|
|||||||
),
|
),
|
||||||
IO.Boolean.Input(
|
IO.Boolean.Input(
|
||||||
"watermark",
|
"watermark",
|
||||||
default=True,
|
default=False,
|
||||||
tooltip='Whether to add an "AI generated" watermark to the image',
|
tooltip='Whether to add an "AI generated" watermark to the image',
|
||||||
optional=True,
|
optional=True,
|
||||||
),
|
),
|
||||||
@ -346,7 +346,7 @@ class ByteDanceSeedreamNode(IO.ComfyNode):
|
|||||||
),
|
),
|
||||||
IO.Boolean.Input(
|
IO.Boolean.Input(
|
||||||
"watermark",
|
"watermark",
|
||||||
default=True,
|
default=False,
|
||||||
tooltip='Whether to add an "AI generated" watermark to the image.',
|
tooltip='Whether to add an "AI generated" watermark to the image.',
|
||||||
optional=True,
|
optional=True,
|
||||||
),
|
),
|
||||||
@ -380,7 +380,7 @@ class ByteDanceSeedreamNode(IO.ComfyNode):
|
|||||||
sequential_image_generation: str = "disabled",
|
sequential_image_generation: str = "disabled",
|
||||||
max_images: int = 1,
|
max_images: int = 1,
|
||||||
seed: int = 0,
|
seed: int = 0,
|
||||||
watermark: bool = True,
|
watermark: bool = False,
|
||||||
fail_on_partial: bool = True,
|
fail_on_partial: bool = True,
|
||||||
) -> IO.NodeOutput:
|
) -> IO.NodeOutput:
|
||||||
validate_string(prompt, strip_whitespace=True, min_length=1)
|
validate_string(prompt, strip_whitespace=True, min_length=1)
|
||||||
@ -507,7 +507,7 @@ class ByteDanceTextToVideoNode(IO.ComfyNode):
|
|||||||
),
|
),
|
||||||
IO.Boolean.Input(
|
IO.Boolean.Input(
|
||||||
"watermark",
|
"watermark",
|
||||||
default=True,
|
default=False,
|
||||||
tooltip='Whether to add an "AI generated" watermark to the video.',
|
tooltip='Whether to add an "AI generated" watermark to the video.',
|
||||||
optional=True,
|
optional=True,
|
||||||
),
|
),
|
||||||
@ -617,7 +617,7 @@ class ByteDanceImageToVideoNode(IO.ComfyNode):
|
|||||||
),
|
),
|
||||||
IO.Boolean.Input(
|
IO.Boolean.Input(
|
||||||
"watermark",
|
"watermark",
|
||||||
default=True,
|
default=False,
|
||||||
tooltip='Whether to add an "AI generated" watermark to the video.',
|
tooltip='Whether to add an "AI generated" watermark to the video.',
|
||||||
optional=True,
|
optional=True,
|
||||||
),
|
),
|
||||||
@ -739,7 +739,7 @@ class ByteDanceFirstLastFrameNode(IO.ComfyNode):
|
|||||||
),
|
),
|
||||||
IO.Boolean.Input(
|
IO.Boolean.Input(
|
||||||
"watermark",
|
"watermark",
|
||||||
default=True,
|
default=False,
|
||||||
tooltip='Whether to add an "AI generated" watermark to the video.',
|
tooltip='Whether to add an "AI generated" watermark to the video.',
|
||||||
optional=True,
|
optional=True,
|
||||||
),
|
),
|
||||||
@ -862,7 +862,7 @@ class ByteDanceImageReferenceNode(IO.ComfyNode):
|
|||||||
),
|
),
|
||||||
IO.Boolean.Input(
|
IO.Boolean.Input(
|
||||||
"watermark",
|
"watermark",
|
||||||
default=True,
|
default=False,
|
||||||
tooltip='Whether to add an "AI generated" watermark to the video.',
|
tooltip='Whether to add an "AI generated" watermark to the video.',
|
||||||
optional=True,
|
optional=True,
|
||||||
),
|
),
|
||||||
|
|||||||
@ -46,14 +46,14 @@ class Txt2ImageParametersField(BaseModel):
|
|||||||
n: int = Field(1, description="Number of images to generate.") # we support only value=1
|
n: int = Field(1, description="Number of images to generate.") # we support only value=1
|
||||||
seed: int = Field(..., ge=0, le=2147483647)
|
seed: int = Field(..., ge=0, le=2147483647)
|
||||||
prompt_extend: bool = Field(True)
|
prompt_extend: bool = Field(True)
|
||||||
watermark: bool = Field(True)
|
watermark: bool = Field(False)
|
||||||
|
|
||||||
|
|
||||||
class Image2ImageParametersField(BaseModel):
|
class Image2ImageParametersField(BaseModel):
|
||||||
size: str | None = Field(None)
|
size: str | None = Field(None)
|
||||||
n: int = Field(1, description="Number of images to generate.") # we support only value=1
|
n: int = Field(1, description="Number of images to generate.") # we support only value=1
|
||||||
seed: int = Field(..., ge=0, le=2147483647)
|
seed: int = Field(..., ge=0, le=2147483647)
|
||||||
watermark: bool = Field(True)
|
watermark: bool = Field(False)
|
||||||
|
|
||||||
|
|
||||||
class Text2VideoParametersField(BaseModel):
|
class Text2VideoParametersField(BaseModel):
|
||||||
@ -61,7 +61,7 @@ class Text2VideoParametersField(BaseModel):
|
|||||||
seed: int = Field(..., ge=0, le=2147483647)
|
seed: int = Field(..., ge=0, le=2147483647)
|
||||||
duration: int = Field(5, ge=5, le=15)
|
duration: int = Field(5, ge=5, le=15)
|
||||||
prompt_extend: bool = Field(True)
|
prompt_extend: bool = Field(True)
|
||||||
watermark: bool = Field(True)
|
watermark: bool = Field(False)
|
||||||
audio: bool = Field(False, description="Whether to generate audio automatically.")
|
audio: bool = Field(False, description="Whether to generate audio automatically.")
|
||||||
shot_type: str = Field("single")
|
shot_type: str = Field("single")
|
||||||
|
|
||||||
@ -71,7 +71,7 @@ class Image2VideoParametersField(BaseModel):
|
|||||||
seed: int = Field(..., ge=0, le=2147483647)
|
seed: int = Field(..., ge=0, le=2147483647)
|
||||||
duration: int = Field(5, ge=5, le=15)
|
duration: int = Field(5, ge=5, le=15)
|
||||||
prompt_extend: bool = Field(True)
|
prompt_extend: bool = Field(True)
|
||||||
watermark: bool = Field(True)
|
watermark: bool = Field(False)
|
||||||
audio: bool = Field(False, description="Whether to generate audio automatically.")
|
audio: bool = Field(False, description="Whether to generate audio automatically.")
|
||||||
shot_type: str = Field("single")
|
shot_type: str = Field("single")
|
||||||
|
|
||||||
@ -208,7 +208,7 @@ class WanTextToImageApi(IO.ComfyNode):
|
|||||||
),
|
),
|
||||||
IO.Boolean.Input(
|
IO.Boolean.Input(
|
||||||
"watermark",
|
"watermark",
|
||||||
default=True,
|
default=False,
|
||||||
tooltip="Whether to add an AI-generated watermark to the result.",
|
tooltip="Whether to add an AI-generated watermark to the result.",
|
||||||
optional=True,
|
optional=True,
|
||||||
),
|
),
|
||||||
@ -234,7 +234,7 @@ class WanTextToImageApi(IO.ComfyNode):
|
|||||||
height: int = 1024,
|
height: int = 1024,
|
||||||
seed: int = 0,
|
seed: int = 0,
|
||||||
prompt_extend: bool = True,
|
prompt_extend: bool = True,
|
||||||
watermark: bool = True,
|
watermark: bool = False,
|
||||||
):
|
):
|
||||||
initial_response = await sync_op(
|
initial_response = await sync_op(
|
||||||
cls,
|
cls,
|
||||||
@ -327,7 +327,7 @@ class WanImageToImageApi(IO.ComfyNode):
|
|||||||
),
|
),
|
||||||
IO.Boolean.Input(
|
IO.Boolean.Input(
|
||||||
"watermark",
|
"watermark",
|
||||||
default=True,
|
default=False,
|
||||||
tooltip="Whether to add an AI-generated watermark to the result.",
|
tooltip="Whether to add an AI-generated watermark to the result.",
|
||||||
optional=True,
|
optional=True,
|
||||||
),
|
),
|
||||||
@ -353,7 +353,7 @@ class WanImageToImageApi(IO.ComfyNode):
|
|||||||
# width: int = 1024,
|
# width: int = 1024,
|
||||||
# height: int = 1024,
|
# height: int = 1024,
|
||||||
seed: int = 0,
|
seed: int = 0,
|
||||||
watermark: bool = True,
|
watermark: bool = False,
|
||||||
):
|
):
|
||||||
n_images = get_number_of_images(image)
|
n_images = get_number_of_images(image)
|
||||||
if n_images not in (1, 2):
|
if n_images not in (1, 2):
|
||||||
@ -476,7 +476,7 @@ class WanTextToVideoApi(IO.ComfyNode):
|
|||||||
),
|
),
|
||||||
IO.Boolean.Input(
|
IO.Boolean.Input(
|
||||||
"watermark",
|
"watermark",
|
||||||
default=True,
|
default=False,
|
||||||
tooltip="Whether to add an AI-generated watermark to the result.",
|
tooltip="Whether to add an AI-generated watermark to the result.",
|
||||||
optional=True,
|
optional=True,
|
||||||
),
|
),
|
||||||
@ -512,7 +512,7 @@ class WanTextToVideoApi(IO.ComfyNode):
|
|||||||
seed: int = 0,
|
seed: int = 0,
|
||||||
generate_audio: bool = False,
|
generate_audio: bool = False,
|
||||||
prompt_extend: bool = True,
|
prompt_extend: bool = True,
|
||||||
watermark: bool = True,
|
watermark: bool = False,
|
||||||
shot_type: str = "single",
|
shot_type: str = "single",
|
||||||
):
|
):
|
||||||
if "480p" in size and model == "wan2.6-t2v":
|
if "480p" in size and model == "wan2.6-t2v":
|
||||||
@ -637,7 +637,7 @@ class WanImageToVideoApi(IO.ComfyNode):
|
|||||||
),
|
),
|
||||||
IO.Boolean.Input(
|
IO.Boolean.Input(
|
||||||
"watermark",
|
"watermark",
|
||||||
default=True,
|
default=False,
|
||||||
tooltip="Whether to add an AI-generated watermark to the result.",
|
tooltip="Whether to add an AI-generated watermark to the result.",
|
||||||
optional=True,
|
optional=True,
|
||||||
),
|
),
|
||||||
@ -674,7 +674,7 @@ class WanImageToVideoApi(IO.ComfyNode):
|
|||||||
seed: int = 0,
|
seed: int = 0,
|
||||||
generate_audio: bool = False,
|
generate_audio: bool = False,
|
||||||
prompt_extend: bool = True,
|
prompt_extend: bool = True,
|
||||||
watermark: bool = True,
|
watermark: bool = False,
|
||||||
shot_type: str = "single",
|
shot_type: str = "single",
|
||||||
):
|
):
|
||||||
if get_number_of_images(image) != 1:
|
if get_number_of_images(image) != 1:
|
||||||
|
|||||||
Loading…
Reference in New Issue
Block a user