mirror of
https://github.com/comfyanonymous/ComfyUI.git
synced 2026-01-10 14:20:49 +08:00
Update README.md
This commit is contained in:
parent
80b967364c
commit
4e91556820
479
README.md
479
README.md
@ -829,205 +829,432 @@ You can pass additional extra model path configurations with one or more copies
|
||||
### Command Line Arguments
|
||||
|
||||
```
|
||||
usage: comfyui [-h] [-c CONFIG_FILE] [--write-out-config-file CONFIG_OUTPUT_PATH] [-w CWD] [--base-paths BASE_PATHS [BASE_PATHS ...]] [-H [IP]] [--port PORT]
|
||||
[--enable-cors-header [ORIGIN]] [--max-upload-size MAX_UPLOAD_SIZE] [--base-directory BASE_DIRECTORY] [--extra-model-paths-config PATH [PATH ...]]
|
||||
[--output-directory OUTPUT_DIRECTORY] [--temp-directory TEMP_DIRECTORY] [--input-directory INPUT_DIRECTORY] [--auto-launch] [--disable-auto-launch]
|
||||
[--cuda-device DEVICE_ID] [--cuda-malloc | --disable-cuda-malloc] [--force-fp32 | --force-fp16 | --force-bf16]
|
||||
[--fp32-unet | --fp64-unet | --bf16-unet | --fp16-unet | --fp8_e4m3fn-unet | --fp8_e5m2-unet] [--fp16-vae | --fp32-vae | --bf16-vae] [--cpu-vae]
|
||||
[--fp8_e4m3fn-text-enc | --fp8_e5m2-text-enc | --fp16-text-enc | --fp32-text-enc | --bf16-text-enc] [--directml [DIRECTML_DEVICE]]
|
||||
[--oneapi-device-selector SELECTOR_STRING] [--disable-ipex-optimize] [--preview-method [none,auto,latent2rgb,taesd]] [--preview-size PREVIEW_SIZE]
|
||||
[--cache-classic | --cache-lru CACHE_LRU | --cache-none]
|
||||
[--use-split-cross-attention | --use-quad-cross-attention | --use-pytorch-cross-attention | --use-sage-attention | --use-flash-attention] [--disable-xformers]
|
||||
[--force-upcast-attention | --dont-upcast-attention] [--gpu-only | --highvram | --normalvram | --lowvram | --novram | --cpu] [--reserve-vram RESERVE_VRAM]
|
||||
[--default-hashing-function {md5,sha1,sha256,sha512}] [--disable-smart-memory] [--deterministic] [--fast [FAST ...]] [--dont-print-server] [--quick-test-for-ci]
|
||||
[--windows-standalone-build] [--disable-metadata] [--disable-all-custom-nodes] [--multi-user] [--create-directories] [--log-stdout]
|
||||
[--plausible-analytics-base-url PLAUSIBLE_ANALYTICS_BASE_URL] [--plausible-analytics-domain PLAUSIBLE_ANALYTICS_DOMAIN] [--analytics-use-identity-provider]
|
||||
[--distributed-queue-connection-uri DISTRIBUTED_QUEUE_CONNECTION_URI] [--distributed-queue-worker] [--distributed-queue-frontend]
|
||||
[--distributed-queue-name DISTRIBUTED_QUEUE_NAME] [--external-address EXTERNAL_ADDRESS] [--logging-level {DEBUG,INFO,WARNING,ERROR,CRITICAL}]
|
||||
[--disable-known-models] [--max-queue-size MAX_QUEUE_SIZE] [--otel-service-name OTEL_SERVICE_NAME] [--otel-service-version OTEL_SERVICE_VERSION]
|
||||
[--otel-exporter-otlp-endpoint OTEL_EXPORTER_OTLP_ENDPOINT] [--force-channels-last] [--force-hf-local-dir-mode] [--front-end-version FRONT_END_VERSION]
|
||||
[--panic-when PANIC_WHEN] [--front-end-root FRONT_END_ROOT] [--executor-factory EXECUTOR_FACTORY] [--openai-api-key OPENAI_API_KEY]
|
||||
[--ideogram-api-key IDEOGRAM_API_KEY] [--anthropic-api-key ANTHROPIC_API_KEY] [--user-directory USER_DIRECTORY] [--enable-compress-response-body]
|
||||
[--workflows WORKFLOWS [WORKFLOWS ...]]
|
||||
usage: comfyui.exe [-h] [-c CONFIG_FILE]
|
||||
[--write-out-config-file CONFIG_OUTPUT_PATH] [-w CWD]
|
||||
[--base-paths BASE_PATHS [BASE_PATHS ...]] [-H [IP]]
|
||||
[--port PORT] [--enable-cors-header [ORIGIN]]
|
||||
[--max-upload-size MAX_UPLOAD_SIZE]
|
||||
[--base-directory BASE_DIRECTORY]
|
||||
[--extra-model-paths-config PATH [PATH ...]]
|
||||
[--output-directory OUTPUT_DIRECTORY]
|
||||
[--temp-directory TEMP_DIRECTORY]
|
||||
[--input-directory INPUT_DIRECTORY] [--auto-launch]
|
||||
[--disable-auto-launch] [--cuda-device DEVICE_ID]
|
||||
[--cuda-malloc | --disable-cuda-malloc]
|
||||
[--force-fp32 | --force-fp16 | --force-bf16]
|
||||
[--fp32-unet | --fp64-unet | --bf16-unet | --fp16-unet | --fp8_e4m3fn-unet | --fp8_e5m2-unet | --fp8_e8m0fnu-unet]
|
||||
[--fp16-vae | --fp32-vae | --bf16-vae] [--cpu-vae]
|
||||
[--fp8_e4m3fn-text-enc | --fp8_e5m2-text-enc | --fp16-text-enc | --fp32-text-enc | --bf16-text-enc]
|
||||
[--directml [DIRECTML_DEVICE]]
|
||||
[--oneapi-device-selector SELECTOR_STRING]
|
||||
[--disable-ipex-optimize] [--supports-fp8-compute]
|
||||
[--preview-method [none,auto,latent2rgb,taesd]]
|
||||
[--preview-size PREVIEW_SIZE]
|
||||
[--cache-classic | --cache-lru CACHE_LRU | --cache-none]
|
||||
[--use-split-cross-attention | --use-quad-cross-attention | --use-pytorch-cross-attention | --use-sage-attention | --use-flash-attention]
|
||||
[--disable-xformers]
|
||||
[--force-upcast-attention | --dont-upcast-attention]
|
||||
[--gpu-only | --highvram | --normalvram | --lowvram | --novram | --cpu]
|
||||
[--reserve-vram RESERVE_VRAM] [--async-offload]
|
||||
[--default-hashing-function {md5,sha1,sha256,sha512}]
|
||||
[--disable-smart-memory] [--deterministic]
|
||||
[--fast [FAST ...]] [--mmap-torch-files]
|
||||
[--dont-print-server] [--quick-test-for-ci]
|
||||
[--windows-standalone-build] [--disable-metadata]
|
||||
[--disable-all-custom-nodes]
|
||||
[--whitelist-custom-nodes WHITELIST_CUSTOM_NODES [WHITELIST_CUSTOM_NODES ...]]
|
||||
[--disable-api-nodes] [--multi-user] [--create-directories]
|
||||
[--log-stdout]
|
||||
[--plausible-analytics-base-url PLAUSIBLE_ANALYTICS_BASE_URL]
|
||||
[--plausible-analytics-domain PLAUSIBLE_ANALYTICS_DOMAIN]
|
||||
[--analytics-use-identity-provider]
|
||||
[--distributed-queue-connection-uri DISTRIBUTED_QUEUE_CONNECTION_URI]
|
||||
[--distributed-queue-worker] [--distributed-queue-frontend]
|
||||
[--distributed-queue-name DISTRIBUTED_QUEUE_NAME]
|
||||
[--external-address EXTERNAL_ADDRESS]
|
||||
[--logging-level {DEBUG,INFO,WARNING,ERROR,CRITICAL}]
|
||||
[--disable-known-models] [--max-queue-size MAX_QUEUE_SIZE]
|
||||
[--otel-service-name OTEL_SERVICE_NAME]
|
||||
[--otel-service-version OTEL_SERVICE_VERSION]
|
||||
[--otel-exporter-otlp-endpoint OTEL_EXPORTER_OTLP_ENDPOINT]
|
||||
[--force-channels-last] [--force-hf-local-dir-mode]
|
||||
[--front-end-version FRONT_END_VERSION]
|
||||
[--panic-when PANIC_WHEN] [--front-end-root FRONT_END_ROOT]
|
||||
[--executor-factory EXECUTOR_FACTORY]
|
||||
[--openai-api-key OPENAI_API_KEY]
|
||||
[--ideogram-api-key IDEOGRAM_API_KEY]
|
||||
[--anthropic-api-key ANTHROPIC_API_KEY]
|
||||
[--user-directory USER_DIRECTORY]
|
||||
[--enable-compress-response-body]
|
||||
[--comfy-api-base COMFY_API_BASE]
|
||||
[--database-url DATABASE_URL]
|
||||
[--workflows WORKFLOWS [WORKFLOWS ...]]
|
||||
[--blip-model-url BLIP_MODEL_URL]
|
||||
[--blip-model-vqa-url BLIP_MODEL_VQA_URL]
|
||||
[--sam-model-vith-url SAM_MODEL_VITH_URL]
|
||||
[--sam-model-vitl-url SAM_MODEL_VITL_URL]
|
||||
[--sam-model-vitb-url SAM_MODEL_VITB_URL]
|
||||
[--history-display-limit HISTORY_DISPLAY_LIMIT]
|
||||
[--ffmpeg-bin-path FFMPEG_BIN_PATH]
|
||||
[--ffmpeg-extra-codecs FFMPEG_EXTRA_CODECS]
|
||||
[--wildcards-path WILDCARDS_PATH]
|
||||
[--wildcard-api WILDCARD_API]
|
||||
[--photoprism-host PHOTOPRISM_HOST]
|
||||
[--immich-host IMMICH_HOST]
|
||||
[--ideogram-session-cookie IDEOGRAM_SESSION_COOKIE]
|
||||
[--use-symlinks] [--ort-providers ORT_PROVIDERS]
|
||||
[--vfi-ops-backend VFI_OPS_BACKEND]
|
||||
[--dependency-version DEPENDENCY_VERSION] [--mmdet-skip]
|
||||
[--sam-editor-cpu] [--sam-editor-model SAM_EDITOR_MODEL]
|
||||
[--custom-wildcards CUSTOM_WILDCARDS]
|
||||
[--disable-gpu-opencv]
|
||||
|
||||
options:
|
||||
-h, --help show this help message and exit
|
||||
-c CONFIG_FILE, --config CONFIG_FILE
|
||||
config file path
|
||||
--write-out-config-file CONFIG_OUTPUT_PATH
|
||||
takes the current command line args and writes them out to a config file at the given path, then exits
|
||||
-w CWD, --cwd CWD Specify the working directory. If not set, this is the current working directory. models/, input/, output/ and other directories will be located here by
|
||||
default. [env var: COMFYUI_CWD]
|
||||
takes the current command line args and writes them
|
||||
out to a config file at the given path, then exits
|
||||
-w CWD, --cwd CWD Specify the working directory. If not set, this is the
|
||||
current working directory. models/, input/, output/
|
||||
and other directories will be located here by default.
|
||||
[env var: COMFYUI_CWD]
|
||||
--base-paths BASE_PATHS [BASE_PATHS ...]
|
||||
Additional base paths for custom nodes, models and inputs. [env var: COMFYUI_BASE_PATHS]
|
||||
Additional base paths for custom nodes, models and
|
||||
inputs. [env var: COMFYUI_BASE_PATHS]
|
||||
-H [IP], --listen [IP]
|
||||
Specify the IP address to listen on (default: 127.0.0.1). You can give a list of ip addresses by separating them with a comma like: 127.2.2.2,127.3.3.3
|
||||
If --listen is provided without an argument, it defaults to 0.0.0.0,:: (listens on all ipv4 and ipv6) [env var: COMFYUI_LISTEN]
|
||||
Specify the IP address to listen on (default:
|
||||
127.0.0.1). You can give a list of ip addresses by
|
||||
separating them with a comma like: 127.2.2.2,127.3.3.3
|
||||
If --listen is provided without an argument, it
|
||||
defaults to 0.0.0.0,:: (listens on all ipv4 and ipv6)
|
||||
[env var: COMFYUI_LISTEN]
|
||||
--port PORT Set the listen port. [env var: COMFYUI_PORT]
|
||||
--enable-cors-header [ORIGIN]
|
||||
Enable CORS (Cross-Origin Resource Sharing) with optional origin or allow all with default '*'. [env var: COMFYUI_ENABLE_CORS_HEADER]
|
||||
Enable CORS (Cross-Origin Resource Sharing) with
|
||||
optional origin or allow all with default '*'. [env
|
||||
var: COMFYUI_ENABLE_CORS_HEADER]
|
||||
--max-upload-size MAX_UPLOAD_SIZE
|
||||
Set the maximum upload size in MB. [env var: COMFYUI_MAX_UPLOAD_SIZE]
|
||||
Set the maximum upload size in MB. [env var:
|
||||
COMFYUI_MAX_UPLOAD_SIZE]
|
||||
--base-directory BASE_DIRECTORY
|
||||
Set the ComfyUI base directory for models, custom_nodes, input, output, temp, and user directories. [env var: COMFYUI_BASE_DIRECTORY]
|
||||
Set the ComfyUI base directory for models,
|
||||
custom_nodes, input, output, temp, and user
|
||||
directories. [env var: COMFYUI_BASE_DIRECTORY]
|
||||
--extra-model-paths-config PATH [PATH ...]
|
||||
Load one or more extra_model_paths.yaml files. [env var: COMFYUI_EXTRA_MODEL_PATHS_CONFIG]
|
||||
Load one or more extra_model_paths.yaml files. [env
|
||||
var: COMFYUI_EXTRA_MODEL_PATHS_CONFIG]
|
||||
--output-directory OUTPUT_DIRECTORY
|
||||
Set the ComfyUI output directory. Overrides --base-directory. [env var: COMFYUI_OUTPUT_DIRECTORY]
|
||||
Set the ComfyUI output directory. Overrides --base-
|
||||
directory. [env var: COMFYUI_OUTPUT_DIRECTORY]
|
||||
--temp-directory TEMP_DIRECTORY
|
||||
Set the ComfyUI temp directory (default is in the ComfyUI directory). Overrides --base-directory. [env var: COMFYUI_TEMP_DIRECTORY]
|
||||
Set the ComfyUI temp directory (default is in the
|
||||
ComfyUI directory). Overrides --base-directory. [env
|
||||
var: COMFYUI_TEMP_DIRECTORY]
|
||||
--input-directory INPUT_DIRECTORY
|
||||
Set the ComfyUI input directory. Overrides --base-directory. [env var: COMFYUI_INPUT_DIRECTORY]
|
||||
--auto-launch Automatically launch ComfyUI in the default browser. [env var: COMFYUI_AUTO_LAUNCH]
|
||||
Set the ComfyUI input directory. Overrides --base-
|
||||
directory. [env var: COMFYUI_INPUT_DIRECTORY]
|
||||
--auto-launch Automatically launch ComfyUI in the default browser.
|
||||
[env var: COMFYUI_AUTO_LAUNCH]
|
||||
--disable-auto-launch
|
||||
Disable auto launching the browser. [env var: COMFYUI_DISABLE_AUTO_LAUNCH]
|
||||
Disable auto launching the browser. [env var:
|
||||
COMFYUI_DISABLE_AUTO_LAUNCH]
|
||||
--cuda-device DEVICE_ID
|
||||
Set the id of the cuda device this instance will use. [env var: COMFYUI_CUDA_DEVICE]
|
||||
--cuda-malloc Enable cudaMallocAsync (enabled by default for torch 2.0 and up). [env var: COMFYUI_CUDA_MALLOC]
|
||||
Set the id of the cuda device this instance will use.
|
||||
[env var: COMFYUI_CUDA_DEVICE]
|
||||
--cuda-malloc Enable cudaMallocAsync (enabled by default for torch
|
||||
2.0 and up). [env var: COMFYUI_CUDA_MALLOC]
|
||||
--disable-cuda-malloc
|
||||
Disable cudaMallocAsync. [env var: COMFYUI_DISABLE_CUDA_MALLOC]
|
||||
--force-fp32 Force fp32 (If this makes your GPU work better please report it). [env var: COMFYUI_FORCE_FP32]
|
||||
Disable cudaMallocAsync. [env var:
|
||||
COMFYUI_DISABLE_CUDA_MALLOC]
|
||||
--force-fp32 Force fp32 (If this makes your GPU work better please
|
||||
report it). [env var: COMFYUI_FORCE_FP32]
|
||||
--force-fp16 Force fp16. [env var: COMFYUI_FORCE_FP16]
|
||||
--force-bf16 Force bf16. [env var: COMFYUI_FORCE_BF16]
|
||||
--fp32-unet Run the diffusion model in fp32. [env var: COMFYUI_FP32_UNET]
|
||||
--fp64-unet Run the diffusion model in fp64. [env var: COMFYUI_FP64_UNET]
|
||||
--bf16-unet Run the diffusion model in bf16. [env var: COMFYUI_BF16_UNET]
|
||||
--fp16-unet Run the diffusion model in fp16 [env var: COMFYUI_FP16_UNET]
|
||||
--fp8_e4m3fn-unet Store unet weights in fp8_e4m3fn. [env var: COMFYUI_FP8_E4M3FN_UNET]
|
||||
--fp8_e5m2-unet Store unet weights in fp8_e5m2. [env var: COMFYUI_FP8_E5M2_UNET]
|
||||
--fp16-vae Run the VAE in fp16, might cause black images. [env var: COMFYUI_FP16_VAE]
|
||||
--fp32-vae Run the VAE in full precision fp32. [env var: COMFYUI_FP32_VAE]
|
||||
--fp32-unet Run the diffusion model in fp32. [env var:
|
||||
COMFYUI_FP32_UNET]
|
||||
--fp64-unet Run the diffusion model in fp64. [env var:
|
||||
COMFYUI_FP64_UNET]
|
||||
--bf16-unet Run the diffusion model in bf16. [env var:
|
||||
COMFYUI_BF16_UNET]
|
||||
--fp16-unet Run the diffusion model in fp16 [env var:
|
||||
COMFYUI_FP16_UNET]
|
||||
--fp8_e4m3fn-unet Store unet weights in fp8_e4m3fn. [env var:
|
||||
COMFYUI_FP8_E4M3FN_UNET]
|
||||
--fp8_e5m2-unet Store unet weights in fp8_e5m2. [env var:
|
||||
COMFYUI_FP8_E5M2_UNET]
|
||||
--fp8_e8m0fnu-unet Store unet weights in fp8_e8m0fnu. [env var:
|
||||
COMFYUI_FP8_E8M0FNU_UNET]
|
||||
--fp16-vae Run the VAE in fp16, might cause black images. [env
|
||||
var: COMFYUI_FP16_VAE]
|
||||
--fp32-vae Run the VAE in full precision fp32. [env var:
|
||||
COMFYUI_FP32_VAE]
|
||||
--bf16-vae Run the VAE in bf16. [env var: COMFYUI_BF16_VAE]
|
||||
--cpu-vae Run the VAE on the CPU. [env var: COMFYUI_CPU_VAE]
|
||||
--fp8_e4m3fn-text-enc
|
||||
Store text encoder weights in fp8 (e4m3fn variant). [env var: COMFYUI_FP8_E4M3FN_TEXT_ENC]
|
||||
--fp8_e5m2-text-enc Store text encoder weights in fp8 (e5m2 variant). [env var: COMFYUI_FP8_E5M2_TEXT_ENC]
|
||||
--fp16-text-enc Store text encoder weights in fp16. [env var: COMFYUI_FP16_TEXT_ENC]
|
||||
--fp32-text-enc Store text encoder weights in fp32. [env var: COMFYUI_FP32_TEXT_ENC]
|
||||
--bf16-text-enc Store text encoder weights in bf16. [env var: COMFYUI_BF16_TEXT_ENC]
|
||||
Store text encoder weights in fp8 (e4m3fn variant).
|
||||
[env var: COMFYUI_FP8_E4M3FN_TEXT_ENC]
|
||||
--fp8_e5m2-text-enc Store text encoder weights in fp8 (e5m2 variant). [env
|
||||
var: COMFYUI_FP8_E5M2_TEXT_ENC]
|
||||
--fp16-text-enc Store text encoder weights in fp16. [env var:
|
||||
COMFYUI_FP16_TEXT_ENC]
|
||||
--fp32-text-enc Store text encoder weights in fp32. [env var:
|
||||
COMFYUI_FP32_TEXT_ENC]
|
||||
--bf16-text-enc Store text encoder weights in bf16. [env var:
|
||||
COMFYUI_BF16_TEXT_ENC]
|
||||
--directml [DIRECTML_DEVICE]
|
||||
Use torch-directml. [env var: COMFYUI_DIRECTML]
|
||||
--oneapi-device-selector SELECTOR_STRING
|
||||
Sets the oneAPI device(s) this instance will use. [env var: COMFYUI_ONEAPI_DEVICE_SELECTOR]
|
||||
Sets the oneAPI device(s) this instance will use. [env
|
||||
var: COMFYUI_ONEAPI_DEVICE_SELECTOR]
|
||||
--disable-ipex-optimize
|
||||
Disables ipex.optimize default when loading models with Intel's Extension for Pytorch. [env var: COMFYUI_DISABLE_IPEX_OPTIMIZE]
|
||||
Disables ipex.optimize default when loading models
|
||||
with Intel's Extension for Pytorch. [env var:
|
||||
COMFYUI_DISABLE_IPEX_OPTIMIZE]
|
||||
--supports-fp8-compute
|
||||
ComfyUI will act like if the device supports fp8
|
||||
compute. [env var: COMFYUI_SUPPORTS_FP8_COMPUTE]
|
||||
--preview-method [none,auto,latent2rgb,taesd]
|
||||
Default preview method for sampler nodes. [env var: COMFYUI_PREVIEW_METHOD]
|
||||
Default preview method for sampler nodes. [env var:
|
||||
COMFYUI_PREVIEW_METHOD]
|
||||
--preview-size PREVIEW_SIZE
|
||||
Sets the maximum preview size for sampler nodes. [env var: COMFYUI_PREVIEW_SIZE]
|
||||
--cache-classic WARNING: Unused. Use the old style (aggressive) caching. [env var: COMFYUI_CACHE_CLASSIC]
|
||||
Sets the maximum preview size for sampler nodes. [env
|
||||
var: COMFYUI_PREVIEW_SIZE]
|
||||
--cache-classic WARNING: Unused. Use the old style (aggressive)
|
||||
caching. [env var: COMFYUI_CACHE_CLASSIC]
|
||||
--cache-lru CACHE_LRU
|
||||
Use LRU caching with a maximum of N node results cached. May use more RAM/VRAM. [env var: COMFYUI_CACHE_LRU]
|
||||
--cache-none Reduced RAM/VRAM usage at the expense of executing every node for each run. [env var: COMFYUI_CACHE_NONE]
|
||||
Use LRU caching with a maximum of N node results
|
||||
cached. May use more RAM/VRAM. [env var:
|
||||
COMFYUI_CACHE_LRU]
|
||||
--cache-none Reduced RAM/VRAM usage at the expense of executing
|
||||
every node for each run. [env var: COMFYUI_CACHE_NONE]
|
||||
--use-split-cross-attention
|
||||
Use the split cross attention optimization. Ignored when xformers is used. [env var: COMFYUI_USE_SPLIT_CROSS_ATTENTION]
|
||||
Use the split cross attention optimization. Ignored
|
||||
when xformers is used. [env var:
|
||||
COMFYUI_USE_SPLIT_CROSS_ATTENTION]
|
||||
--use-quad-cross-attention
|
||||
Use the sub-quadratic cross attention optimization . Ignored when xformers is used. [env var: COMFYUI_USE_QUAD_CROSS_ATTENTION]
|
||||
Use the sub-quadratic cross attention optimization .
|
||||
Ignored when xformers is used. [env var:
|
||||
COMFYUI_USE_QUAD_CROSS_ATTENTION]
|
||||
--use-pytorch-cross-attention
|
||||
Use the new pytorch 2.0 cross attention function (default). [env var: COMFYUI_USE_PYTORCH_CROSS_ATTENTION]
|
||||
--use-sage-attention Use sage attention. [env var: COMFYUI_USE_SAGE_ATTENTION]
|
||||
Use the new pytorch 2.0 cross attention function
|
||||
(default). [env var:
|
||||
COMFYUI_USE_PYTORCH_CROSS_ATTENTION]
|
||||
--use-sage-attention Use sage attention. [env var:
|
||||
COMFYUI_USE_SAGE_ATTENTION]
|
||||
--use-flash-attention
|
||||
Use FlashAttention. [env var: COMFYUI_USE_FLASH_ATTENTION]
|
||||
Use FlashAttention. [env var:
|
||||
COMFYUI_USE_FLASH_ATTENTION]
|
||||
--disable-xformers Disable xformers. [env var: COMFYUI_DISABLE_XFORMERS]
|
||||
--force-upcast-attention
|
||||
Force enable attention upcasting, please report if it fixes black images. [env var: COMFYUI_FORCE_UPCAST_ATTENTION]
|
||||
Force enable attention upcasting, please report if it
|
||||
fixes black images. [env var:
|
||||
COMFYUI_FORCE_UPCAST_ATTENTION]
|
||||
--dont-upcast-attention
|
||||
Disable all upcasting of attention. Should be unnecessary except for debugging. [env var: COMFYUI_DONT_UPCAST_ATTENTION]
|
||||
--gpu-only Store and run everything (text encoders/CLIP models, etc... on the GPU). [env var: COMFYUI_GPU_ONLY]
|
||||
--highvram By default models will be unloaded to CPU memory after being used. This option keeps them in GPU memory. [env var: COMFYUI_HIGHVRAM]
|
||||
--normalvram Used to force normal vram use if lowvram gets automatically enabled. [env var: COMFYUI_NORMALVRAM]
|
||||
--lowvram Split the unet in parts to use less vram. [env var: COMFYUI_LOWVRAM]
|
||||
Disable all upcasting of attention. Should be
|
||||
unnecessary except for debugging. [env var:
|
||||
COMFYUI_DONT_UPCAST_ATTENTION]
|
||||
--gpu-only Store and run everything (text encoders/CLIP models,
|
||||
etc... on the GPU). [env var: COMFYUI_GPU_ONLY]
|
||||
--highvram By default models will be unloaded to CPU memory after
|
||||
being used. This option keeps them in GPU memory. [env
|
||||
var: COMFYUI_HIGHVRAM]
|
||||
--normalvram Used to force normal vram use if lowvram gets
|
||||
automatically enabled. [env var: COMFYUI_NORMALVRAM]
|
||||
--lowvram Split the unet in parts to use less vram. [env var:
|
||||
COMFYUI_LOWVRAM]
|
||||
--novram When lowvram isn't enough. [env var: COMFYUI_NOVRAM]
|
||||
--cpu To use the CPU for everything (slow). [env var: COMFYUI_CPU]
|
||||
--cpu To use the CPU for everything (slow). [env var:
|
||||
COMFYUI_CPU]
|
||||
--reserve-vram RESERVE_VRAM
|
||||
Set the amount of vram in GB you want to reserve for use by your OS/other software. By default some amount is reserved depending on your OS. [env var:
|
||||
Set the amount of vram in GB you want to reserve for
|
||||
use by your OS/other software. By default some amount
|
||||
is reserved depending on your OS. [env var:
|
||||
COMFYUI_RESERVE_VRAM]
|
||||
--async-offload Use async weight offloading. [env var:
|
||||
COMFYUI_ASYNC_OFFLOAD]
|
||||
--default-hashing-function {md5,sha1,sha256,sha512}
|
||||
Allows you to choose the hash function to use for duplicate filename / contents comparison. Default is sha256. [env var:
|
||||
COMFYUI_DEFAULT_HASHING_FUNCTION]
|
||||
Allows you to choose the hash function to use for
|
||||
duplicate filename / contents comparison. Default is
|
||||
sha256. [env var: COMFYUI_DEFAULT_HASHING_FUNCTION]
|
||||
--disable-smart-memory
|
||||
Force ComfyUI to agressively offload to regular ram instead of keeping models in vram when it can. [env var: COMFYUI_DISABLE_SMART_MEMORY]
|
||||
--deterministic Make pytorch use slower deterministic algorithms when it can. Note that this might not make images deterministic in all cases. [env var:
|
||||
Force ComfyUI to agressively offload to regular ram
|
||||
instead of keeping models in vram when it can. [env
|
||||
var: COMFYUI_DISABLE_SMART_MEMORY]
|
||||
--deterministic Make pytorch use slower deterministic algorithms when
|
||||
it can. Note that this might not make images
|
||||
deterministic in all cases. [env var:
|
||||
COMFYUI_DETERMINISTIC]
|
||||
--fast [FAST ...] Enable some untested and potentially quality deteriorating optimizations. Pass a list specific optimizations if you only want to enable specific ones.
|
||||
Current valid optimizations: fp16_accumulation fp8_matrix_mult cublas_ops [env var: COMFYUI_FAST]
|
||||
--dont-print-server Don't print server output. [env var: COMFYUI_DONT_PRINT_SERVER]
|
||||
--quick-test-for-ci Quick test for CI. Raises an error if nodes cannot be imported, [env var: COMFYUI_QUICK_TEST_FOR_CI]
|
||||
--fast [FAST ...] Enable some untested and potentially quality
|
||||
deteriorating optimizations. Pass a list specific
|
||||
optimizations if you only want to enable specific
|
||||
ones. Current valid optimizations: fp16_accumulation
|
||||
fp8_matrix_mult cublas_ops [env var: COMFYUI_FAST]
|
||||
--mmap-torch-files Use mmap when loading ckpt/pt files. [env var:
|
||||
COMFYUI_MMAP_TORCH_FILES]
|
||||
--dont-print-server Don't print server output. [env var:
|
||||
COMFYUI_DONT_PRINT_SERVER]
|
||||
--quick-test-for-ci Quick test for CI. Raises an error if nodes cannot be
|
||||
imported, [env var: COMFYUI_QUICK_TEST_FOR_CI]
|
||||
--windows-standalone-build
|
||||
Windows standalone build: Enable convenient things that most people using the standalone windows build will probably enjoy (like auto opening the page
|
||||
on startup). [env var: COMFYUI_WINDOWS_STANDALONE_BUILD]
|
||||
--disable-metadata Disable saving prompt metadata in files. [env var: COMFYUI_DISABLE_METADATA]
|
||||
Windows standalone build: Enable convenient things
|
||||
that most people using the standalone windows build
|
||||
will probably enjoy (like auto opening the page on
|
||||
startup). [env var: COMFYUI_WINDOWS_STANDALONE_BUILD]
|
||||
--disable-metadata Disable saving prompt metadata in files. [env var:
|
||||
COMFYUI_DISABLE_METADATA]
|
||||
--disable-all-custom-nodes
|
||||
Disable loading all custom nodes. [env var: COMFYUI_DISABLE_ALL_CUSTOM_NODES]
|
||||
--multi-user Enables per-user storage. [env var: COMFYUI_MULTI_USER]
|
||||
--create-directories Creates the default models/, input/, output/ and temp/ directories, then exits. [env var: COMFYUI_CREATE_DIRECTORIES]
|
||||
--log-stdout Send normal process output to stdout instead of stderr (default). [env var: COMFYUI_LOG_STDOUT]
|
||||
Disable loading all custom nodes. [env var:
|
||||
COMFYUI_DISABLE_ALL_CUSTOM_NODES]
|
||||
--whitelist-custom-nodes WHITELIST_CUSTOM_NODES [WHITELIST_CUSTOM_NODES ...]
|
||||
Specify custom node folders to load even when
|
||||
--disable-all-custom-nodes is enabled. [env var:
|
||||
COMFYUI_WHITELIST_CUSTOM_NODES]
|
||||
--disable-api-nodes Disable loading all api nodes. [env var:
|
||||
COMFYUI_DISABLE_API_NODES]
|
||||
--multi-user Enables per-user storage. [env var:
|
||||
COMFYUI_MULTI_USER]
|
||||
--create-directories Creates the default models/, input/, output/ and temp/
|
||||
directories, then exits. [env var:
|
||||
COMFYUI_CREATE_DIRECTORIES]
|
||||
--log-stdout Send normal process output to stdout instead of stderr
|
||||
(default). [env var: COMFYUI_LOG_STDOUT]
|
||||
--plausible-analytics-base-url PLAUSIBLE_ANALYTICS_BASE_URL
|
||||
Enables server-side analytics events sent to the provided URL. [env var: COMFYUI_PLAUSIBLE_ANALYTICS_BASE_URL]
|
||||
Enables server-side analytics events sent to the
|
||||
provided URL. [env var:
|
||||
COMFYUI_PLAUSIBLE_ANALYTICS_BASE_URL]
|
||||
--plausible-analytics-domain PLAUSIBLE_ANALYTICS_DOMAIN
|
||||
Specifies the domain name for analytics events. [env var: COMFYUI_PLAUSIBLE_ANALYTICS_DOMAIN]
|
||||
Specifies the domain name for analytics events. [env
|
||||
var: COMFYUI_PLAUSIBLE_ANALYTICS_DOMAIN]
|
||||
--analytics-use-identity-provider
|
||||
Uses platform identifiers for unique visitor analytics. [env var: COMFYUI_ANALYTICS_USE_IDENTITY_PROVIDER]
|
||||
Uses platform identifiers for unique visitor
|
||||
analytics. [env var:
|
||||
COMFYUI_ANALYTICS_USE_IDENTITY_PROVIDER]
|
||||
--distributed-queue-connection-uri DISTRIBUTED_QUEUE_CONNECTION_URI
|
||||
EXAMPLE: "amqp://guest:guest@127.0.0.1" - Servers and clients will connect to this AMPQ URL to form a distributed queue and exchange prompt execution
|
||||
requests and progress updates. [env var: COMFYUI_DISTRIBUTED_QUEUE_CONNECTION_URI]
|
||||
EXAMPLE: "amqp://guest:guest@127.0.0.1" - Servers and
|
||||
clients will connect to this AMPQ URL to form a
|
||||
distributed queue and exchange prompt execution
|
||||
requests and progress updates. [env var:
|
||||
COMFYUI_DISTRIBUTED_QUEUE_CONNECTION_URI]
|
||||
--distributed-queue-worker
|
||||
Workers will pull requests off the AMQP URL. [env var: COMFYUI_DISTRIBUTED_QUEUE_WORKER]
|
||||
Workers will pull requests off the AMQP URL. [env var:
|
||||
COMFYUI_DISTRIBUTED_QUEUE_WORKER]
|
||||
--distributed-queue-frontend
|
||||
Frontends will start the web UI and connect to the provided AMQP URL to submit prompts. [env var: COMFYUI_DISTRIBUTED_QUEUE_FRONTEND]
|
||||
Frontends will start the web UI and connect to the
|
||||
provided AMQP URL to submit prompts. [env var:
|
||||
COMFYUI_DISTRIBUTED_QUEUE_FRONTEND]
|
||||
--distributed-queue-name DISTRIBUTED_QUEUE_NAME
|
||||
This name will be used by the frontends and workers to exchange prompt requests and replies. Progress updates will be prefixed by the queue name,
|
||||
followed by a '.', then the user ID [env var: COMFYUI_DISTRIBUTED_QUEUE_NAME]
|
||||
This name will be used by the frontends and workers to
|
||||
exchange prompt requests and replies. Progress updates
|
||||
will be prefixed by the queue name, followed by a '.',
|
||||
then the user ID [env var:
|
||||
COMFYUI_DISTRIBUTED_QUEUE_NAME]
|
||||
--external-address EXTERNAL_ADDRESS
|
||||
Specifies a base URL for external addresses reported by the API, such as for image paths. [env var: COMFYUI_EXTERNAL_ADDRESS]
|
||||
Specifies a base URL for external addresses reported
|
||||
by the API, such as for image paths. [env var:
|
||||
COMFYUI_EXTERNAL_ADDRESS]
|
||||
--logging-level {DEBUG,INFO,WARNING,ERROR,CRITICAL}
|
||||
Set the logging level [env var: COMFYUI_LOGGING_LEVEL]
|
||||
--disable-known-models
|
||||
Disables automatic downloads of known models and prevents them from appearing in the UI. [env var: COMFYUI_DISABLE_KNOWN_MODELS]
|
||||
Disables automatic downloads of known models and
|
||||
prevents them from appearing in the UI. [env var:
|
||||
COMFYUI_DISABLE_KNOWN_MODELS]
|
||||
--max-queue-size MAX_QUEUE_SIZE
|
||||
The API will reject prompt requests if the queue's size exceeds this value. [env var: COMFYUI_MAX_QUEUE_SIZE]
|
||||
The API will reject prompt requests if the queue's
|
||||
size exceeds this value. [env var:
|
||||
COMFYUI_MAX_QUEUE_SIZE]
|
||||
--otel-service-name OTEL_SERVICE_NAME
|
||||
The name of the service or application that is generating telemetry data. [env var: OTEL_SERVICE_NAME]
|
||||
The name of the service or application that is
|
||||
generating telemetry data. [env var:
|
||||
OTEL_SERVICE_NAME]
|
||||
--otel-service-version OTEL_SERVICE_VERSION
|
||||
The version of the service or application that is generating telemetry data. [env var: OTEL_SERVICE_VERSION]
|
||||
The version of the service or application that is
|
||||
generating telemetry data. [env var:
|
||||
OTEL_SERVICE_VERSION]
|
||||
--otel-exporter-otlp-endpoint OTEL_EXPORTER_OTLP_ENDPOINT
|
||||
A base endpoint URL for any signal type, with an optionally-specified port number. Helpful for when you're sending more than one signal to the same
|
||||
endpoint and want one environment variable to control the endpoint. [env var: OTEL_EXPORTER_OTLP_ENDPOINT]
|
||||
A base endpoint URL for any signal type, with an
|
||||
optionally-specified port number. Helpful for when
|
||||
you're sending more than one signal to the same
|
||||
endpoint and want one environment variable to control
|
||||
the endpoint. [env var: OTEL_EXPORTER_OTLP_ENDPOINT]
|
||||
--force-channels-last
|
||||
Force channels last format when inferencing the models. [env var: COMFYUI_FORCE_CHANNELS_LAST]
|
||||
Force channels last format when inferencing the
|
||||
models. [env var: COMFYUI_FORCE_CHANNELS_LAST]
|
||||
--force-hf-local-dir-mode
|
||||
Download repos from huggingface.co to the models/huggingface directory with the "local_dir" argument instead of models/huggingface_cache with the
|
||||
"cache_dir" argument, recreating the traditional file structure. [env var: COMFYUI_FORCE_HF_LOCAL_DIR_MODE]
|
||||
Download repos from huggingface.co to the
|
||||
models/huggingface directory with the "local_dir"
|
||||
argument instead of models/huggingface_cache with the
|
||||
"cache_dir" argument, recreating the traditional file
|
||||
structure. [env var: COMFYUI_FORCE_HF_LOCAL_DIR_MODE]
|
||||
--front-end-version FRONT_END_VERSION
|
||||
Specifies the version of the frontend to be used. This command needs internet connectivity to query and download available frontend implementations from
|
||||
GitHub releases. The version string should be in the format of: [repoOwner]/[repoName]@[version] where version is one of: "latest" or a valid version
|
||||
number (e.g. "1.0.0") [env var: COMFYUI_FRONT_END_VERSION]
|
||||
Specifies the version of the frontend to be used. This
|
||||
command needs internet connectivity to query and
|
||||
download available frontend implementations from
|
||||
GitHub releases. The version string should be in the
|
||||
format of: [repoOwner]/[repoName]@[version] where
|
||||
version is one of: "latest" or a valid version number
|
||||
(e.g. "1.0.0") [env var: COMFYUI_FRONT_END_VERSION]
|
||||
--panic-when PANIC_WHEN
|
||||
List of fully qualified exception class names to panic (sys.exit(1)) when a workflow raises it. Example: --panic-when=torch.cuda.OutOfMemoryError. Can
|
||||
be specified multiple times or as a comma-separated list. [env var: COMFYUI_PANIC_WHEN]
|
||||
List of fully qualified exception class names to panic
|
||||
(sys.exit(1)) when a workflow raises it. Example:
|
||||
--panic-when=torch.cuda.OutOfMemoryError. Can be
|
||||
specified multiple times or as a comma-separated list.
|
||||
[env var: COMFYUI_PANIC_WHEN]
|
||||
--front-end-root FRONT_END_ROOT
|
||||
The local filesystem path to the directory where the frontend is located. Overrides --front-end-version. [env var: COMFYUI_FRONT_END_ROOT]
|
||||
The local filesystem path to the directory where the
|
||||
frontend is located. Overrides --front-end-version.
|
||||
[env var: COMFYUI_FRONT_END_ROOT]
|
||||
--executor-factory EXECUTOR_FACTORY
|
||||
When running ComfyUI as a distributed worker, this specifies the kind of executor that should be used to run the actual ComfyUI workflow worker. A
|
||||
ThreadPoolExecutor is the default. A ProcessPoolExecutor results in better memory management, since the process will be closed and large, contiguous
|
||||
blocks of CUDA memory can be freed. [env var: COMFYUI_EXECUTOR_FACTORY]
|
||||
When running ComfyUI as a distributed worker, this
|
||||
specifies the kind of executor that should be used to
|
||||
run the actual ComfyUI workflow worker. A
|
||||
ThreadPoolExecutor is the default. A
|
||||
ProcessPoolExecutor results in better memory
|
||||
management, since the process will be closed and
|
||||
large, contiguous blocks of CUDA memory can be freed.
|
||||
[env var: COMFYUI_EXECUTOR_FACTORY]
|
||||
--openai-api-key OPENAI_API_KEY
|
||||
Configures the OpenAI API Key for the OpenAI nodes. Visit https://platform.openai.com/api-keys to create this key. [env var: OPENAI_API_KEY]
|
||||
Configures the OpenAI API Key for the OpenAI nodes.
|
||||
Visit https://platform.openai.com/api-keys to create
|
||||
this key. [env var: OPENAI_API_KEY]
|
||||
--ideogram-api-key IDEOGRAM_API_KEY
|
||||
Configures the Ideogram API Key for the Ideogram nodes. Visit https://ideogram.ai/manage-api to create this key. [env var: IDEOGRAM_API_KEY]
|
||||
Configures the Ideogram API Key for the Ideogram
|
||||
nodes. Visit https://ideogram.ai/manage-api to create
|
||||
this key. [env var: IDEOGRAM_API_KEY]
|
||||
--anthropic-api-key ANTHROPIC_API_KEY
|
||||
Configures the Anthropic API key for its nodes related to Claude functionality. Visit https://console.anthropic.com/settings/keys to create this key.
|
||||
[env var: ANTHROPIC_API_KEY]
|
||||
Configures the Anthropic API key for its nodes related
|
||||
to Claude functionality. Visit
|
||||
https://console.anthropic.com/settings/keys to create
|
||||
this key. [env var: ANTHROPIC_API_KEY]
|
||||
--user-directory USER_DIRECTORY
|
||||
Set the ComfyUI user directory with an absolute path. Overrides --base-directory. [env var: COMFYUI_USER_DIRECTORY]
|
||||
Set the ComfyUI user directory with an absolute path.
|
||||
Overrides --base-directory. [env var:
|
||||
COMFYUI_USER_DIRECTORY]
|
||||
--enable-compress-response-body
|
||||
Enable compressing response body. [env var: COMFYUI_ENABLE_COMPRESS_RESPONSE_BODY]
|
||||
Enable compressing response body. [env var:
|
||||
COMFYUI_ENABLE_COMPRESS_RESPONSE_BODY]
|
||||
--comfy-api-base COMFY_API_BASE
|
||||
Set the base URL for the ComfyUI API. (default:
|
||||
https://api.comfy.org) [env var:
|
||||
COMFYUI_COMFY_API_BASE]
|
||||
--database-url DATABASE_URL
|
||||
Specify the database URL, e.g. for an in-memory
|
||||
database you can use 'sqlite:///:memory:'. [env var:
|
||||
COMFYUI_DATABASE_URL]
|
||||
--workflows WORKFLOWS [WORKFLOWS ...]
|
||||
Execute the API workflow(s) specified in the provided files. For each workflow, its outputs will be printed to a line to standard out. Application
|
||||
logging will be redirected to standard error. Use `-` to signify standard in. [env var: COMFYUI_WORKFLOWS]
|
||||
|
||||
Args that start with '--' can also be set in a config file (config.yaml or config.json or specified via -c). Config file syntax allows: key=value, flag=true, stuff=[a,b,c] (for details, see syntax at
|
||||
https://goo.gl/R74nmi). In general, command-line values override environment variables which override config file values which override defaults.
|
||||
Execute the API workflow(s) specified in the provided
|
||||
files. For each workflow, its outputs will be printed
|
||||
to a line to standard out. Application logging will be
|
||||
redirected to standard error. Use `-` to signify
|
||||
standard in. [env var: COMFYUI_WORKFLOWS]
|
||||
|
||||
Args that start with '--' can also be set in a config file (config.yaml or
|
||||
config.json or specified via -c). Config file syntax allows: key=value,
|
||||
flag=true, stuff=[a,b,c] (for details, see syntax at https://goo.gl/R74nmi).
|
||||
In general, command-line values override environment variables which override
|
||||
config file values which override defaults.
|
||||
|
||||
```
|
||||
|
||||
# Using ComfyUI as an API / Programmatically
|
||||
|
||||
Loading…
Reference in New Issue
Block a user