Update README.md

This commit is contained in:
doctorpangloss 2025-07-15 10:50:34 -07:00
parent 80b967364c
commit 4e91556820

477
README.md
View File

@ -829,205 +829,432 @@ You can pass additional extra model path configurations with one or more copies
### Command Line Arguments ### Command Line Arguments
``` ```
usage: comfyui [-h] [-c CONFIG_FILE] [--write-out-config-file CONFIG_OUTPUT_PATH] [-w CWD] [--base-paths BASE_PATHS [BASE_PATHS ...]] [-H [IP]] [--port PORT] usage: comfyui.exe [-h] [-c CONFIG_FILE]
[--enable-cors-header [ORIGIN]] [--max-upload-size MAX_UPLOAD_SIZE] [--base-directory BASE_DIRECTORY] [--extra-model-paths-config PATH [PATH ...]] [--write-out-config-file CONFIG_OUTPUT_PATH] [-w CWD]
[--output-directory OUTPUT_DIRECTORY] [--temp-directory TEMP_DIRECTORY] [--input-directory INPUT_DIRECTORY] [--auto-launch] [--disable-auto-launch] [--base-paths BASE_PATHS [BASE_PATHS ...]] [-H [IP]]
[--cuda-device DEVICE_ID] [--cuda-malloc | --disable-cuda-malloc] [--force-fp32 | --force-fp16 | --force-bf16] [--port PORT] [--enable-cors-header [ORIGIN]]
[--fp32-unet | --fp64-unet | --bf16-unet | --fp16-unet | --fp8_e4m3fn-unet | --fp8_e5m2-unet] [--fp16-vae | --fp32-vae | --bf16-vae] [--cpu-vae] [--max-upload-size MAX_UPLOAD_SIZE]
[--fp8_e4m3fn-text-enc | --fp8_e5m2-text-enc | --fp16-text-enc | --fp32-text-enc | --bf16-text-enc] [--directml [DIRECTML_DEVICE]] [--base-directory BASE_DIRECTORY]
[--oneapi-device-selector SELECTOR_STRING] [--disable-ipex-optimize] [--preview-method [none,auto,latent2rgb,taesd]] [--preview-size PREVIEW_SIZE] [--extra-model-paths-config PATH [PATH ...]]
[--cache-classic | --cache-lru CACHE_LRU | --cache-none] [--output-directory OUTPUT_DIRECTORY]
[--use-split-cross-attention | --use-quad-cross-attention | --use-pytorch-cross-attention | --use-sage-attention | --use-flash-attention] [--disable-xformers] [--temp-directory TEMP_DIRECTORY]
[--force-upcast-attention | --dont-upcast-attention] [--gpu-only | --highvram | --normalvram | --lowvram | --novram | --cpu] [--reserve-vram RESERVE_VRAM] [--input-directory INPUT_DIRECTORY] [--auto-launch]
[--default-hashing-function {md5,sha1,sha256,sha512}] [--disable-smart-memory] [--deterministic] [--fast [FAST ...]] [--dont-print-server] [--quick-test-for-ci] [--disable-auto-launch] [--cuda-device DEVICE_ID]
[--windows-standalone-build] [--disable-metadata] [--disable-all-custom-nodes] [--multi-user] [--create-directories] [--log-stdout] [--cuda-malloc | --disable-cuda-malloc]
[--plausible-analytics-base-url PLAUSIBLE_ANALYTICS_BASE_URL] [--plausible-analytics-domain PLAUSIBLE_ANALYTICS_DOMAIN] [--analytics-use-identity-provider] [--force-fp32 | --force-fp16 | --force-bf16]
[--distributed-queue-connection-uri DISTRIBUTED_QUEUE_CONNECTION_URI] [--distributed-queue-worker] [--distributed-queue-frontend] [--fp32-unet | --fp64-unet | --bf16-unet | --fp16-unet | --fp8_e4m3fn-unet | --fp8_e5m2-unet | --fp8_e8m0fnu-unet]
[--distributed-queue-name DISTRIBUTED_QUEUE_NAME] [--external-address EXTERNAL_ADDRESS] [--logging-level {DEBUG,INFO,WARNING,ERROR,CRITICAL}] [--fp16-vae | --fp32-vae | --bf16-vae] [--cpu-vae]
[--disable-known-models] [--max-queue-size MAX_QUEUE_SIZE] [--otel-service-name OTEL_SERVICE_NAME] [--otel-service-version OTEL_SERVICE_VERSION] [--fp8_e4m3fn-text-enc | --fp8_e5m2-text-enc | --fp16-text-enc | --fp32-text-enc | --bf16-text-enc]
[--otel-exporter-otlp-endpoint OTEL_EXPORTER_OTLP_ENDPOINT] [--force-channels-last] [--force-hf-local-dir-mode] [--front-end-version FRONT_END_VERSION] [--directml [DIRECTML_DEVICE]]
[--panic-when PANIC_WHEN] [--front-end-root FRONT_END_ROOT] [--executor-factory EXECUTOR_FACTORY] [--openai-api-key OPENAI_API_KEY] [--oneapi-device-selector SELECTOR_STRING]
[--ideogram-api-key IDEOGRAM_API_KEY] [--anthropic-api-key ANTHROPIC_API_KEY] [--user-directory USER_DIRECTORY] [--enable-compress-response-body] [--disable-ipex-optimize] [--supports-fp8-compute]
[--workflows WORKFLOWS [WORKFLOWS ...]] [--preview-method [none,auto,latent2rgb,taesd]]
[--preview-size PREVIEW_SIZE]
[--cache-classic | --cache-lru CACHE_LRU | --cache-none]
[--use-split-cross-attention | --use-quad-cross-attention | --use-pytorch-cross-attention | --use-sage-attention | --use-flash-attention]
[--disable-xformers]
[--force-upcast-attention | --dont-upcast-attention]
[--gpu-only | --highvram | --normalvram | --lowvram | --novram | --cpu]
[--reserve-vram RESERVE_VRAM] [--async-offload]
[--default-hashing-function {md5,sha1,sha256,sha512}]
[--disable-smart-memory] [--deterministic]
[--fast [FAST ...]] [--mmap-torch-files]
[--dont-print-server] [--quick-test-for-ci]
[--windows-standalone-build] [--disable-metadata]
[--disable-all-custom-nodes]
[--whitelist-custom-nodes WHITELIST_CUSTOM_NODES [WHITELIST_CUSTOM_NODES ...]]
[--disable-api-nodes] [--multi-user] [--create-directories]
[--log-stdout]
[--plausible-analytics-base-url PLAUSIBLE_ANALYTICS_BASE_URL]
[--plausible-analytics-domain PLAUSIBLE_ANALYTICS_DOMAIN]
[--analytics-use-identity-provider]
[--distributed-queue-connection-uri DISTRIBUTED_QUEUE_CONNECTION_URI]
[--distributed-queue-worker] [--distributed-queue-frontend]
[--distributed-queue-name DISTRIBUTED_QUEUE_NAME]
[--external-address EXTERNAL_ADDRESS]
[--logging-level {DEBUG,INFO,WARNING,ERROR,CRITICAL}]
[--disable-known-models] [--max-queue-size MAX_QUEUE_SIZE]
[--otel-service-name OTEL_SERVICE_NAME]
[--otel-service-version OTEL_SERVICE_VERSION]
[--otel-exporter-otlp-endpoint OTEL_EXPORTER_OTLP_ENDPOINT]
[--force-channels-last] [--force-hf-local-dir-mode]
[--front-end-version FRONT_END_VERSION]
[--panic-when PANIC_WHEN] [--front-end-root FRONT_END_ROOT]
[--executor-factory EXECUTOR_FACTORY]
[--openai-api-key OPENAI_API_KEY]
[--ideogram-api-key IDEOGRAM_API_KEY]
[--anthropic-api-key ANTHROPIC_API_KEY]
[--user-directory USER_DIRECTORY]
[--enable-compress-response-body]
[--comfy-api-base COMFY_API_BASE]
[--database-url DATABASE_URL]
[--workflows WORKFLOWS [WORKFLOWS ...]]
[--blip-model-url BLIP_MODEL_URL]
[--blip-model-vqa-url BLIP_MODEL_VQA_URL]
[--sam-model-vith-url SAM_MODEL_VITH_URL]
[--sam-model-vitl-url SAM_MODEL_VITL_URL]
[--sam-model-vitb-url SAM_MODEL_VITB_URL]
[--history-display-limit HISTORY_DISPLAY_LIMIT]
[--ffmpeg-bin-path FFMPEG_BIN_PATH]
[--ffmpeg-extra-codecs FFMPEG_EXTRA_CODECS]
[--wildcards-path WILDCARDS_PATH]
[--wildcard-api WILDCARD_API]
[--photoprism-host PHOTOPRISM_HOST]
[--immich-host IMMICH_HOST]
[--ideogram-session-cookie IDEOGRAM_SESSION_COOKIE]
[--use-symlinks] [--ort-providers ORT_PROVIDERS]
[--vfi-ops-backend VFI_OPS_BACKEND]
[--dependency-version DEPENDENCY_VERSION] [--mmdet-skip]
[--sam-editor-cpu] [--sam-editor-model SAM_EDITOR_MODEL]
[--custom-wildcards CUSTOM_WILDCARDS]
[--disable-gpu-opencv]
options: options:
-h, --help show this help message and exit -h, --help show this help message and exit
-c CONFIG_FILE, --config CONFIG_FILE -c CONFIG_FILE, --config CONFIG_FILE
config file path config file path
--write-out-config-file CONFIG_OUTPUT_PATH --write-out-config-file CONFIG_OUTPUT_PATH
takes the current command line args and writes them out to a config file at the given path, then exits takes the current command line args and writes them
-w CWD, --cwd CWD Specify the working directory. If not set, this is the current working directory. models/, input/, output/ and other directories will be located here by out to a config file at the given path, then exits
default. [env var: COMFYUI_CWD] -w CWD, --cwd CWD Specify the working directory. If not set, this is the
current working directory. models/, input/, output/
and other directories will be located here by default.
[env var: COMFYUI_CWD]
--base-paths BASE_PATHS [BASE_PATHS ...] --base-paths BASE_PATHS [BASE_PATHS ...]
Additional base paths for custom nodes, models and inputs. [env var: COMFYUI_BASE_PATHS] Additional base paths for custom nodes, models and
inputs. [env var: COMFYUI_BASE_PATHS]
-H [IP], --listen [IP] -H [IP], --listen [IP]
Specify the IP address to listen on (default: 127.0.0.1). You can give a list of ip addresses by separating them with a comma like: 127.2.2.2,127.3.3.3 Specify the IP address to listen on (default:
If --listen is provided without an argument, it defaults to 0.0.0.0,:: (listens on all ipv4 and ipv6) [env var: COMFYUI_LISTEN] 127.0.0.1). You can give a list of ip addresses by
separating them with a comma like: 127.2.2.2,127.3.3.3
If --listen is provided without an argument, it
defaults to 0.0.0.0,:: (listens on all ipv4 and ipv6)
[env var: COMFYUI_LISTEN]
--port PORT Set the listen port. [env var: COMFYUI_PORT] --port PORT Set the listen port. [env var: COMFYUI_PORT]
--enable-cors-header [ORIGIN] --enable-cors-header [ORIGIN]
Enable CORS (Cross-Origin Resource Sharing) with optional origin or allow all with default '*'. [env var: COMFYUI_ENABLE_CORS_HEADER] Enable CORS (Cross-Origin Resource Sharing) with
optional origin or allow all with default '*'. [env
var: COMFYUI_ENABLE_CORS_HEADER]
--max-upload-size MAX_UPLOAD_SIZE --max-upload-size MAX_UPLOAD_SIZE
Set the maximum upload size in MB. [env var: COMFYUI_MAX_UPLOAD_SIZE] Set the maximum upload size in MB. [env var:
COMFYUI_MAX_UPLOAD_SIZE]
--base-directory BASE_DIRECTORY --base-directory BASE_DIRECTORY
Set the ComfyUI base directory for models, custom_nodes, input, output, temp, and user directories. [env var: COMFYUI_BASE_DIRECTORY] Set the ComfyUI base directory for models,
custom_nodes, input, output, temp, and user
directories. [env var: COMFYUI_BASE_DIRECTORY]
--extra-model-paths-config PATH [PATH ...] --extra-model-paths-config PATH [PATH ...]
Load one or more extra_model_paths.yaml files. [env var: COMFYUI_EXTRA_MODEL_PATHS_CONFIG] Load one or more extra_model_paths.yaml files. [env
var: COMFYUI_EXTRA_MODEL_PATHS_CONFIG]
--output-directory OUTPUT_DIRECTORY --output-directory OUTPUT_DIRECTORY
Set the ComfyUI output directory. Overrides --base-directory. [env var: COMFYUI_OUTPUT_DIRECTORY] Set the ComfyUI output directory. Overrides --base-
directory. [env var: COMFYUI_OUTPUT_DIRECTORY]
--temp-directory TEMP_DIRECTORY --temp-directory TEMP_DIRECTORY
Set the ComfyUI temp directory (default is in the ComfyUI directory). Overrides --base-directory. [env var: COMFYUI_TEMP_DIRECTORY] Set the ComfyUI temp directory (default is in the
ComfyUI directory). Overrides --base-directory. [env
var: COMFYUI_TEMP_DIRECTORY]
--input-directory INPUT_DIRECTORY --input-directory INPUT_DIRECTORY
Set the ComfyUI input directory. Overrides --base-directory. [env var: COMFYUI_INPUT_DIRECTORY] Set the ComfyUI input directory. Overrides --base-
--auto-launch Automatically launch ComfyUI in the default browser. [env var: COMFYUI_AUTO_LAUNCH] directory. [env var: COMFYUI_INPUT_DIRECTORY]
--auto-launch Automatically launch ComfyUI in the default browser.
[env var: COMFYUI_AUTO_LAUNCH]
--disable-auto-launch --disable-auto-launch
Disable auto launching the browser. [env var: COMFYUI_DISABLE_AUTO_LAUNCH] Disable auto launching the browser. [env var:
COMFYUI_DISABLE_AUTO_LAUNCH]
--cuda-device DEVICE_ID --cuda-device DEVICE_ID
Set the id of the cuda device this instance will use. [env var: COMFYUI_CUDA_DEVICE] Set the id of the cuda device this instance will use.
--cuda-malloc Enable cudaMallocAsync (enabled by default for torch 2.0 and up). [env var: COMFYUI_CUDA_MALLOC] [env var: COMFYUI_CUDA_DEVICE]
--cuda-malloc Enable cudaMallocAsync (enabled by default for torch
2.0 and up). [env var: COMFYUI_CUDA_MALLOC]
--disable-cuda-malloc --disable-cuda-malloc
Disable cudaMallocAsync. [env var: COMFYUI_DISABLE_CUDA_MALLOC] Disable cudaMallocAsync. [env var:
--force-fp32 Force fp32 (If this makes your GPU work better please report it). [env var: COMFYUI_FORCE_FP32] COMFYUI_DISABLE_CUDA_MALLOC]
--force-fp32 Force fp32 (If this makes your GPU work better please
report it). [env var: COMFYUI_FORCE_FP32]
--force-fp16 Force fp16. [env var: COMFYUI_FORCE_FP16] --force-fp16 Force fp16. [env var: COMFYUI_FORCE_FP16]
--force-bf16 Force bf16. [env var: COMFYUI_FORCE_BF16] --force-bf16 Force bf16. [env var: COMFYUI_FORCE_BF16]
--fp32-unet Run the diffusion model in fp32. [env var: COMFYUI_FP32_UNET] --fp32-unet Run the diffusion model in fp32. [env var:
--fp64-unet Run the diffusion model in fp64. [env var: COMFYUI_FP64_UNET] COMFYUI_FP32_UNET]
--bf16-unet Run the diffusion model in bf16. [env var: COMFYUI_BF16_UNET] --fp64-unet Run the diffusion model in fp64. [env var:
--fp16-unet Run the diffusion model in fp16 [env var: COMFYUI_FP16_UNET] COMFYUI_FP64_UNET]
--fp8_e4m3fn-unet Store unet weights in fp8_e4m3fn. [env var: COMFYUI_FP8_E4M3FN_UNET] --bf16-unet Run the diffusion model in bf16. [env var:
--fp8_e5m2-unet Store unet weights in fp8_e5m2. [env var: COMFYUI_FP8_E5M2_UNET] COMFYUI_BF16_UNET]
--fp16-vae Run the VAE in fp16, might cause black images. [env var: COMFYUI_FP16_VAE] --fp16-unet Run the diffusion model in fp16 [env var:
--fp32-vae Run the VAE in full precision fp32. [env var: COMFYUI_FP32_VAE] COMFYUI_FP16_UNET]
--fp8_e4m3fn-unet Store unet weights in fp8_e4m3fn. [env var:
COMFYUI_FP8_E4M3FN_UNET]
--fp8_e5m2-unet Store unet weights in fp8_e5m2. [env var:
COMFYUI_FP8_E5M2_UNET]
--fp8_e8m0fnu-unet Store unet weights in fp8_e8m0fnu. [env var:
COMFYUI_FP8_E8M0FNU_UNET]
--fp16-vae Run the VAE in fp16, might cause black images. [env
var: COMFYUI_FP16_VAE]
--fp32-vae Run the VAE in full precision fp32. [env var:
COMFYUI_FP32_VAE]
--bf16-vae Run the VAE in bf16. [env var: COMFYUI_BF16_VAE] --bf16-vae Run the VAE in bf16. [env var: COMFYUI_BF16_VAE]
--cpu-vae Run the VAE on the CPU. [env var: COMFYUI_CPU_VAE] --cpu-vae Run the VAE on the CPU. [env var: COMFYUI_CPU_VAE]
--fp8_e4m3fn-text-enc --fp8_e4m3fn-text-enc
Store text encoder weights in fp8 (e4m3fn variant). [env var: COMFYUI_FP8_E4M3FN_TEXT_ENC] Store text encoder weights in fp8 (e4m3fn variant).
--fp8_e5m2-text-enc Store text encoder weights in fp8 (e5m2 variant). [env var: COMFYUI_FP8_E5M2_TEXT_ENC] [env var: COMFYUI_FP8_E4M3FN_TEXT_ENC]
--fp16-text-enc Store text encoder weights in fp16. [env var: COMFYUI_FP16_TEXT_ENC] --fp8_e5m2-text-enc Store text encoder weights in fp8 (e5m2 variant). [env
--fp32-text-enc Store text encoder weights in fp32. [env var: COMFYUI_FP32_TEXT_ENC] var: COMFYUI_FP8_E5M2_TEXT_ENC]
--bf16-text-enc Store text encoder weights in bf16. [env var: COMFYUI_BF16_TEXT_ENC] --fp16-text-enc Store text encoder weights in fp16. [env var:
COMFYUI_FP16_TEXT_ENC]
--fp32-text-enc Store text encoder weights in fp32. [env var:
COMFYUI_FP32_TEXT_ENC]
--bf16-text-enc Store text encoder weights in bf16. [env var:
COMFYUI_BF16_TEXT_ENC]
--directml [DIRECTML_DEVICE] --directml [DIRECTML_DEVICE]
Use torch-directml. [env var: COMFYUI_DIRECTML] Use torch-directml. [env var: COMFYUI_DIRECTML]
--oneapi-device-selector SELECTOR_STRING --oneapi-device-selector SELECTOR_STRING
Sets the oneAPI device(s) this instance will use. [env var: COMFYUI_ONEAPI_DEVICE_SELECTOR] Sets the oneAPI device(s) this instance will use. [env
var: COMFYUI_ONEAPI_DEVICE_SELECTOR]
--disable-ipex-optimize --disable-ipex-optimize
Disables ipex.optimize default when loading models with Intel's Extension for Pytorch. [env var: COMFYUI_DISABLE_IPEX_OPTIMIZE] Disables ipex.optimize default when loading models
with Intel's Extension for Pytorch. [env var:
COMFYUI_DISABLE_IPEX_OPTIMIZE]
--supports-fp8-compute
ComfyUI will act like if the device supports fp8
compute. [env var: COMFYUI_SUPPORTS_FP8_COMPUTE]
--preview-method [none,auto,latent2rgb,taesd] --preview-method [none,auto,latent2rgb,taesd]
Default preview method for sampler nodes. [env var: COMFYUI_PREVIEW_METHOD] Default preview method for sampler nodes. [env var:
COMFYUI_PREVIEW_METHOD]
--preview-size PREVIEW_SIZE --preview-size PREVIEW_SIZE
Sets the maximum preview size for sampler nodes. [env var: COMFYUI_PREVIEW_SIZE] Sets the maximum preview size for sampler nodes. [env
--cache-classic WARNING: Unused. Use the old style (aggressive) caching. [env var: COMFYUI_CACHE_CLASSIC] var: COMFYUI_PREVIEW_SIZE]
--cache-classic WARNING: Unused. Use the old style (aggressive)
caching. [env var: COMFYUI_CACHE_CLASSIC]
--cache-lru CACHE_LRU --cache-lru CACHE_LRU
Use LRU caching with a maximum of N node results cached. May use more RAM/VRAM. [env var: COMFYUI_CACHE_LRU] Use LRU caching with a maximum of N node results
--cache-none Reduced RAM/VRAM usage at the expense of executing every node for each run. [env var: COMFYUI_CACHE_NONE] cached. May use more RAM/VRAM. [env var:
COMFYUI_CACHE_LRU]
--cache-none Reduced RAM/VRAM usage at the expense of executing
every node for each run. [env var: COMFYUI_CACHE_NONE]
--use-split-cross-attention --use-split-cross-attention
Use the split cross attention optimization. Ignored when xformers is used. [env var: COMFYUI_USE_SPLIT_CROSS_ATTENTION] Use the split cross attention optimization. Ignored
when xformers is used. [env var:
COMFYUI_USE_SPLIT_CROSS_ATTENTION]
--use-quad-cross-attention --use-quad-cross-attention
Use the sub-quadratic cross attention optimization . Ignored when xformers is used. [env var: COMFYUI_USE_QUAD_CROSS_ATTENTION] Use the sub-quadratic cross attention optimization .
Ignored when xformers is used. [env var:
COMFYUI_USE_QUAD_CROSS_ATTENTION]
--use-pytorch-cross-attention --use-pytorch-cross-attention
Use the new pytorch 2.0 cross attention function (default). [env var: COMFYUI_USE_PYTORCH_CROSS_ATTENTION] Use the new pytorch 2.0 cross attention function
--use-sage-attention Use sage attention. [env var: COMFYUI_USE_SAGE_ATTENTION] (default). [env var:
COMFYUI_USE_PYTORCH_CROSS_ATTENTION]
--use-sage-attention Use sage attention. [env var:
COMFYUI_USE_SAGE_ATTENTION]
--use-flash-attention --use-flash-attention
Use FlashAttention. [env var: COMFYUI_USE_FLASH_ATTENTION] Use FlashAttention. [env var:
COMFYUI_USE_FLASH_ATTENTION]
--disable-xformers Disable xformers. [env var: COMFYUI_DISABLE_XFORMERS] --disable-xformers Disable xformers. [env var: COMFYUI_DISABLE_XFORMERS]
--force-upcast-attention --force-upcast-attention
Force enable attention upcasting, please report if it fixes black images. [env var: COMFYUI_FORCE_UPCAST_ATTENTION] Force enable attention upcasting, please report if it
fixes black images. [env var:
COMFYUI_FORCE_UPCAST_ATTENTION]
--dont-upcast-attention --dont-upcast-attention
Disable all upcasting of attention. Should be unnecessary except for debugging. [env var: COMFYUI_DONT_UPCAST_ATTENTION] Disable all upcasting of attention. Should be
--gpu-only Store and run everything (text encoders/CLIP models, etc... on the GPU). [env var: COMFYUI_GPU_ONLY] unnecessary except for debugging. [env var:
--highvram By default models will be unloaded to CPU memory after being used. This option keeps them in GPU memory. [env var: COMFYUI_HIGHVRAM] COMFYUI_DONT_UPCAST_ATTENTION]
--normalvram Used to force normal vram use if lowvram gets automatically enabled. [env var: COMFYUI_NORMALVRAM] --gpu-only Store and run everything (text encoders/CLIP models,
--lowvram Split the unet in parts to use less vram. [env var: COMFYUI_LOWVRAM] etc... on the GPU). [env var: COMFYUI_GPU_ONLY]
--highvram By default models will be unloaded to CPU memory after
being used. This option keeps them in GPU memory. [env
var: COMFYUI_HIGHVRAM]
--normalvram Used to force normal vram use if lowvram gets
automatically enabled. [env var: COMFYUI_NORMALVRAM]
--lowvram Split the unet in parts to use less vram. [env var:
COMFYUI_LOWVRAM]
--novram When lowvram isn't enough. [env var: COMFYUI_NOVRAM] --novram When lowvram isn't enough. [env var: COMFYUI_NOVRAM]
--cpu To use the CPU for everything (slow). [env var: COMFYUI_CPU] --cpu To use the CPU for everything (slow). [env var:
COMFYUI_CPU]
--reserve-vram RESERVE_VRAM --reserve-vram RESERVE_VRAM
Set the amount of vram in GB you want to reserve for use by your OS/other software. By default some amount is reserved depending on your OS. [env var: Set the amount of vram in GB you want to reserve for
use by your OS/other software. By default some amount
is reserved depending on your OS. [env var:
COMFYUI_RESERVE_VRAM] COMFYUI_RESERVE_VRAM]
--async-offload Use async weight offloading. [env var:
COMFYUI_ASYNC_OFFLOAD]
--default-hashing-function {md5,sha1,sha256,sha512} --default-hashing-function {md5,sha1,sha256,sha512}
Allows you to choose the hash function to use for duplicate filename / contents comparison. Default is sha256. [env var: Allows you to choose the hash function to use for
COMFYUI_DEFAULT_HASHING_FUNCTION] duplicate filename / contents comparison. Default is
sha256. [env var: COMFYUI_DEFAULT_HASHING_FUNCTION]
--disable-smart-memory --disable-smart-memory
Force ComfyUI to agressively offload to regular ram instead of keeping models in vram when it can. [env var: COMFYUI_DISABLE_SMART_MEMORY] Force ComfyUI to agressively offload to regular ram
--deterministic Make pytorch use slower deterministic algorithms when it can. Note that this might not make images deterministic in all cases. [env var: instead of keeping models in vram when it can. [env
var: COMFYUI_DISABLE_SMART_MEMORY]
--deterministic Make pytorch use slower deterministic algorithms when
it can. Note that this might not make images
deterministic in all cases. [env var:
COMFYUI_DETERMINISTIC] COMFYUI_DETERMINISTIC]
--fast [FAST ...] Enable some untested and potentially quality deteriorating optimizations. Pass a list specific optimizations if you only want to enable specific ones. --fast [FAST ...] Enable some untested and potentially quality
Current valid optimizations: fp16_accumulation fp8_matrix_mult cublas_ops [env var: COMFYUI_FAST] deteriorating optimizations. Pass a list specific
--dont-print-server Don't print server output. [env var: COMFYUI_DONT_PRINT_SERVER] optimizations if you only want to enable specific
--quick-test-for-ci Quick test for CI. Raises an error if nodes cannot be imported, [env var: COMFYUI_QUICK_TEST_FOR_CI] ones. Current valid optimizations: fp16_accumulation
fp8_matrix_mult cublas_ops [env var: COMFYUI_FAST]
--mmap-torch-files Use mmap when loading ckpt/pt files. [env var:
COMFYUI_MMAP_TORCH_FILES]
--dont-print-server Don't print server output. [env var:
COMFYUI_DONT_PRINT_SERVER]
--quick-test-for-ci Quick test for CI. Raises an error if nodes cannot be
imported, [env var: COMFYUI_QUICK_TEST_FOR_CI]
--windows-standalone-build --windows-standalone-build
Windows standalone build: Enable convenient things that most people using the standalone windows build will probably enjoy (like auto opening the page Windows standalone build: Enable convenient things
on startup). [env var: COMFYUI_WINDOWS_STANDALONE_BUILD] that most people using the standalone windows build
--disable-metadata Disable saving prompt metadata in files. [env var: COMFYUI_DISABLE_METADATA] will probably enjoy (like auto opening the page on
startup). [env var: COMFYUI_WINDOWS_STANDALONE_BUILD]
--disable-metadata Disable saving prompt metadata in files. [env var:
COMFYUI_DISABLE_METADATA]
--disable-all-custom-nodes --disable-all-custom-nodes
Disable loading all custom nodes. [env var: COMFYUI_DISABLE_ALL_CUSTOM_NODES] Disable loading all custom nodes. [env var:
--multi-user Enables per-user storage. [env var: COMFYUI_MULTI_USER] COMFYUI_DISABLE_ALL_CUSTOM_NODES]
--create-directories Creates the default models/, input/, output/ and temp/ directories, then exits. [env var: COMFYUI_CREATE_DIRECTORIES] --whitelist-custom-nodes WHITELIST_CUSTOM_NODES [WHITELIST_CUSTOM_NODES ...]
--log-stdout Send normal process output to stdout instead of stderr (default). [env var: COMFYUI_LOG_STDOUT] Specify custom node folders to load even when
--disable-all-custom-nodes is enabled. [env var:
COMFYUI_WHITELIST_CUSTOM_NODES]
--disable-api-nodes Disable loading all api nodes. [env var:
COMFYUI_DISABLE_API_NODES]
--multi-user Enables per-user storage. [env var:
COMFYUI_MULTI_USER]
--create-directories Creates the default models/, input/, output/ and temp/
directories, then exits. [env var:
COMFYUI_CREATE_DIRECTORIES]
--log-stdout Send normal process output to stdout instead of stderr
(default). [env var: COMFYUI_LOG_STDOUT]
--plausible-analytics-base-url PLAUSIBLE_ANALYTICS_BASE_URL --plausible-analytics-base-url PLAUSIBLE_ANALYTICS_BASE_URL
Enables server-side analytics events sent to the provided URL. [env var: COMFYUI_PLAUSIBLE_ANALYTICS_BASE_URL] Enables server-side analytics events sent to the
provided URL. [env var:
COMFYUI_PLAUSIBLE_ANALYTICS_BASE_URL]
--plausible-analytics-domain PLAUSIBLE_ANALYTICS_DOMAIN --plausible-analytics-domain PLAUSIBLE_ANALYTICS_DOMAIN
Specifies the domain name for analytics events. [env var: COMFYUI_PLAUSIBLE_ANALYTICS_DOMAIN] Specifies the domain name for analytics events. [env
var: COMFYUI_PLAUSIBLE_ANALYTICS_DOMAIN]
--analytics-use-identity-provider --analytics-use-identity-provider
Uses platform identifiers for unique visitor analytics. [env var: COMFYUI_ANALYTICS_USE_IDENTITY_PROVIDER] Uses platform identifiers for unique visitor
analytics. [env var:
COMFYUI_ANALYTICS_USE_IDENTITY_PROVIDER]
--distributed-queue-connection-uri DISTRIBUTED_QUEUE_CONNECTION_URI --distributed-queue-connection-uri DISTRIBUTED_QUEUE_CONNECTION_URI
EXAMPLE: "amqp://guest:guest@127.0.0.1" - Servers and clients will connect to this AMPQ URL to form a distributed queue and exchange prompt execution EXAMPLE: "amqp://guest:guest@127.0.0.1" - Servers and
requests and progress updates. [env var: COMFYUI_DISTRIBUTED_QUEUE_CONNECTION_URI] clients will connect to this AMPQ URL to form a
distributed queue and exchange prompt execution
requests and progress updates. [env var:
COMFYUI_DISTRIBUTED_QUEUE_CONNECTION_URI]
--distributed-queue-worker --distributed-queue-worker
Workers will pull requests off the AMQP URL. [env var: COMFYUI_DISTRIBUTED_QUEUE_WORKER] Workers will pull requests off the AMQP URL. [env var:
COMFYUI_DISTRIBUTED_QUEUE_WORKER]
--distributed-queue-frontend --distributed-queue-frontend
Frontends will start the web UI and connect to the provided AMQP URL to submit prompts. [env var: COMFYUI_DISTRIBUTED_QUEUE_FRONTEND] Frontends will start the web UI and connect to the
provided AMQP URL to submit prompts. [env var:
COMFYUI_DISTRIBUTED_QUEUE_FRONTEND]
--distributed-queue-name DISTRIBUTED_QUEUE_NAME --distributed-queue-name DISTRIBUTED_QUEUE_NAME
This name will be used by the frontends and workers to exchange prompt requests and replies. Progress updates will be prefixed by the queue name, This name will be used by the frontends and workers to
followed by a '.', then the user ID [env var: COMFYUI_DISTRIBUTED_QUEUE_NAME] exchange prompt requests and replies. Progress updates
will be prefixed by the queue name, followed by a '.',
then the user ID [env var:
COMFYUI_DISTRIBUTED_QUEUE_NAME]
--external-address EXTERNAL_ADDRESS --external-address EXTERNAL_ADDRESS
Specifies a base URL for external addresses reported by the API, such as for image paths. [env var: COMFYUI_EXTERNAL_ADDRESS] Specifies a base URL for external addresses reported
by the API, such as for image paths. [env var:
COMFYUI_EXTERNAL_ADDRESS]
--logging-level {DEBUG,INFO,WARNING,ERROR,CRITICAL} --logging-level {DEBUG,INFO,WARNING,ERROR,CRITICAL}
Set the logging level [env var: COMFYUI_LOGGING_LEVEL] Set the logging level [env var: COMFYUI_LOGGING_LEVEL]
--disable-known-models --disable-known-models
Disables automatic downloads of known models and prevents them from appearing in the UI. [env var: COMFYUI_DISABLE_KNOWN_MODELS] Disables automatic downloads of known models and
prevents them from appearing in the UI. [env var:
COMFYUI_DISABLE_KNOWN_MODELS]
--max-queue-size MAX_QUEUE_SIZE --max-queue-size MAX_QUEUE_SIZE
The API will reject prompt requests if the queue's size exceeds this value. [env var: COMFYUI_MAX_QUEUE_SIZE] The API will reject prompt requests if the queue's
size exceeds this value. [env var:
COMFYUI_MAX_QUEUE_SIZE]
--otel-service-name OTEL_SERVICE_NAME --otel-service-name OTEL_SERVICE_NAME
The name of the service or application that is generating telemetry data. [env var: OTEL_SERVICE_NAME] The name of the service or application that is
generating telemetry data. [env var:
OTEL_SERVICE_NAME]
--otel-service-version OTEL_SERVICE_VERSION --otel-service-version OTEL_SERVICE_VERSION
The version of the service or application that is generating telemetry data. [env var: OTEL_SERVICE_VERSION] The version of the service or application that is
generating telemetry data. [env var:
OTEL_SERVICE_VERSION]
--otel-exporter-otlp-endpoint OTEL_EXPORTER_OTLP_ENDPOINT --otel-exporter-otlp-endpoint OTEL_EXPORTER_OTLP_ENDPOINT
A base endpoint URL for any signal type, with an optionally-specified port number. Helpful for when you're sending more than one signal to the same A base endpoint URL for any signal type, with an
endpoint and want one environment variable to control the endpoint. [env var: OTEL_EXPORTER_OTLP_ENDPOINT] optionally-specified port number. Helpful for when
you're sending more than one signal to the same
endpoint and want one environment variable to control
the endpoint. [env var: OTEL_EXPORTER_OTLP_ENDPOINT]
--force-channels-last --force-channels-last
Force channels last format when inferencing the models. [env var: COMFYUI_FORCE_CHANNELS_LAST] Force channels last format when inferencing the
models. [env var: COMFYUI_FORCE_CHANNELS_LAST]
--force-hf-local-dir-mode --force-hf-local-dir-mode
Download repos from huggingface.co to the models/huggingface directory with the "local_dir" argument instead of models/huggingface_cache with the Download repos from huggingface.co to the
"cache_dir" argument, recreating the traditional file structure. [env var: COMFYUI_FORCE_HF_LOCAL_DIR_MODE] models/huggingface directory with the "local_dir"
argument instead of models/huggingface_cache with the
"cache_dir" argument, recreating the traditional file
structure. [env var: COMFYUI_FORCE_HF_LOCAL_DIR_MODE]
--front-end-version FRONT_END_VERSION --front-end-version FRONT_END_VERSION
Specifies the version of the frontend to be used. This command needs internet connectivity to query and download available frontend implementations from Specifies the version of the frontend to be used. This
GitHub releases. The version string should be in the format of: [repoOwner]/[repoName]@[version] where version is one of: "latest" or a valid version command needs internet connectivity to query and
number (e.g. "1.0.0") [env var: COMFYUI_FRONT_END_VERSION] download available frontend implementations from
GitHub releases. The version string should be in the
format of: [repoOwner]/[repoName]@[version] where
version is one of: "latest" or a valid version number
(e.g. "1.0.0") [env var: COMFYUI_FRONT_END_VERSION]
--panic-when PANIC_WHEN --panic-when PANIC_WHEN
List of fully qualified exception class names to panic (sys.exit(1)) when a workflow raises it. Example: --panic-when=torch.cuda.OutOfMemoryError. Can List of fully qualified exception class names to panic
be specified multiple times or as a comma-separated list. [env var: COMFYUI_PANIC_WHEN] (sys.exit(1)) when a workflow raises it. Example:
--panic-when=torch.cuda.OutOfMemoryError. Can be
specified multiple times or as a comma-separated list.
[env var: COMFYUI_PANIC_WHEN]
--front-end-root FRONT_END_ROOT --front-end-root FRONT_END_ROOT
The local filesystem path to the directory where the frontend is located. Overrides --front-end-version. [env var: COMFYUI_FRONT_END_ROOT] The local filesystem path to the directory where the
frontend is located. Overrides --front-end-version.
[env var: COMFYUI_FRONT_END_ROOT]
--executor-factory EXECUTOR_FACTORY --executor-factory EXECUTOR_FACTORY
When running ComfyUI as a distributed worker, this specifies the kind of executor that should be used to run the actual ComfyUI workflow worker. A When running ComfyUI as a distributed worker, this
ThreadPoolExecutor is the default. A ProcessPoolExecutor results in better memory management, since the process will be closed and large, contiguous specifies the kind of executor that should be used to
blocks of CUDA memory can be freed. [env var: COMFYUI_EXECUTOR_FACTORY] run the actual ComfyUI workflow worker. A
ThreadPoolExecutor is the default. A
ProcessPoolExecutor results in better memory
management, since the process will be closed and
large, contiguous blocks of CUDA memory can be freed.
[env var: COMFYUI_EXECUTOR_FACTORY]
--openai-api-key OPENAI_API_KEY --openai-api-key OPENAI_API_KEY
Configures the OpenAI API Key for the OpenAI nodes. Visit https://platform.openai.com/api-keys to create this key. [env var: OPENAI_API_KEY] Configures the OpenAI API Key for the OpenAI nodes.
Visit https://platform.openai.com/api-keys to create
this key. [env var: OPENAI_API_KEY]
--ideogram-api-key IDEOGRAM_API_KEY --ideogram-api-key IDEOGRAM_API_KEY
Configures the Ideogram API Key for the Ideogram nodes. Visit https://ideogram.ai/manage-api to create this key. [env var: IDEOGRAM_API_KEY] Configures the Ideogram API Key for the Ideogram
nodes. Visit https://ideogram.ai/manage-api to create
this key. [env var: IDEOGRAM_API_KEY]
--anthropic-api-key ANTHROPIC_API_KEY --anthropic-api-key ANTHROPIC_API_KEY
Configures the Anthropic API key for its nodes related to Claude functionality. Visit https://console.anthropic.com/settings/keys to create this key. Configures the Anthropic API key for its nodes related
[env var: ANTHROPIC_API_KEY] to Claude functionality. Visit
https://console.anthropic.com/settings/keys to create
this key. [env var: ANTHROPIC_API_KEY]
--user-directory USER_DIRECTORY --user-directory USER_DIRECTORY
Set the ComfyUI user directory with an absolute path. Overrides --base-directory. [env var: COMFYUI_USER_DIRECTORY] Set the ComfyUI user directory with an absolute path.
Overrides --base-directory. [env var:
COMFYUI_USER_DIRECTORY]
--enable-compress-response-body --enable-compress-response-body
Enable compressing response body. [env var: COMFYUI_ENABLE_COMPRESS_RESPONSE_BODY] Enable compressing response body. [env var:
COMFYUI_ENABLE_COMPRESS_RESPONSE_BODY]
--comfy-api-base COMFY_API_BASE
Set the base URL for the ComfyUI API. (default:
https://api.comfy.org) [env var:
COMFYUI_COMFY_API_BASE]
--database-url DATABASE_URL
Specify the database URL, e.g. for an in-memory
database you can use 'sqlite:///:memory:'. [env var:
COMFYUI_DATABASE_URL]
--workflows WORKFLOWS [WORKFLOWS ...] --workflows WORKFLOWS [WORKFLOWS ...]
Execute the API workflow(s) specified in the provided files. For each workflow, its outputs will be printed to a line to standard out. Application Execute the API workflow(s) specified in the provided
logging will be redirected to standard error. Use `-` to signify standard in. [env var: COMFYUI_WORKFLOWS] files. For each workflow, its outputs will be printed
to a line to standard out. Application logging will be
redirected to standard error. Use `-` to signify
standard in. [env var: COMFYUI_WORKFLOWS]
Args that start with '--' can also be set in a config file (config.yaml or
config.json or specified via -c). Config file syntax allows: key=value,
flag=true, stuff=[a,b,c] (for details, see syntax at https://goo.gl/R74nmi).
In general, command-line values override environment variables which override
config file values which override defaults.
Args that start with '--' can also be set in a config file (config.yaml or config.json or specified via -c). Config file syntax allows: key=value, flag=true, stuff=[a,b,c] (for details, see syntax at
https://goo.gl/R74nmi). In general, command-line values override environment variables which override config file values which override defaults.
``` ```
# Using ComfyUI as an API / Programmatically # Using ComfyUI as an API / Programmatically