mirror of
https://github.com/comfyanonymous/ComfyUI.git
synced 2026-01-11 23:00:51 +08:00
Fix cross attention docs
This commit is contained in:
parent
1c22b82e09
commit
340974dec3
@ -302,8 +302,8 @@ CONFIG_OPTIONS = [
|
|||||||
OptionInfo("directml", type=int, nargs="?", metavar="DIRECTML_DEVICE", const=-1,
|
OptionInfo("directml", type=int, nargs="?", metavar="DIRECTML_DEVICE", const=-1,
|
||||||
help="Use torch-directml."),
|
help="Use torch-directml."),
|
||||||
OptionInfoEnum("cross-attention", [
|
OptionInfoEnum("cross-attention", [
|
||||||
OptionInfoEnumChoice("split", option_name="use-split-cross-attention", help="By default models will be unloaded to CPU memory after being used. This option keeps them in GPU memory."),
|
OptionInfoEnumChoice("split", option_name="use-split-cross-attention", help="Use the split cross attention optimization instead of the sub-quadratic one. Ignored when xformers is used."),
|
||||||
OptionInfoEnumChoice("pytorch", option_name="use-pytorch-cross-attention", help="Used to force normal vram use if lowvram gets automatically enabled."),
|
OptionInfoEnumChoice("pytorch", option_name="use-pytorch-cross-attention", help="Use the new pytorch 2.0 cross attention function."),
|
||||||
], help="Type of cross attention to use", empty_help="Don't use cross-attention."),
|
], help="Type of cross attention to use", empty_help="Don't use cross-attention."),
|
||||||
OptionInfoFlag("disable-xformers",
|
OptionInfoFlag("disable-xformers",
|
||||||
help="Disable xformers."),
|
help="Disable xformers."),
|
||||||
|
|||||||
16
config.yaml
16
config.yaml
@ -17,17 +17,17 @@ config:
|
|||||||
# Extra paths to scan for model files.
|
# Extra paths to scan for model files.
|
||||||
extra_model_paths:
|
extra_model_paths:
|
||||||
a1111:
|
a1111:
|
||||||
checkpoints: models/Stable-diffusion
|
|
||||||
vae: models/VAE
|
|
||||||
hypernetworks: models/hypernetworks
|
hypernetworks: models/hypernetworks
|
||||||
base_path: path/to/stable-diffusion-webui/
|
base_path: path/to/stable-diffusion-webui/
|
||||||
upscale_models: |
|
embeddings: embeddings
|
||||||
models/ESRGAN
|
|
||||||
models/SwinIR
|
|
||||||
controlnet: models/ControlNet
|
controlnet: models/ControlNet
|
||||||
configs: models/Stable-diffusion
|
configs: models/Stable-diffusion
|
||||||
loras: models/Lora
|
loras: models/Lora
|
||||||
embeddings: embeddings
|
vae: models/VAE
|
||||||
|
checkpoints: models/Stable-diffusion
|
||||||
|
upscale_models: |
|
||||||
|
models/ESRGAN
|
||||||
|
models/SwinIR
|
||||||
|
|
||||||
# Set the ComfyUI output directory. Leave empty to use the default.
|
# Set the ComfyUI output directory. Leave empty to use the default.
|
||||||
# Type: str
|
# Type: str
|
||||||
@ -70,8 +70,8 @@ config:
|
|||||||
# Type of cross attention to use
|
# Type of cross attention to use
|
||||||
# Choices:
|
# Choices:
|
||||||
# - (empty): Don't use cross-attention.
|
# - (empty): Don't use cross-attention.
|
||||||
# - split: By default models will be unloaded to CPU memory after being used. This option keeps them in GPU memory.
|
# - split: Use the split cross attention optimization instead of the sub-quadratic one. Ignored when xformers is used.
|
||||||
# - pytorch: Used to force normal vram use if lowvram gets automatically enabled.
|
# - pytorch: Use the new pytorch 2.0 cross attention function.
|
||||||
cross_attention:
|
cross_attention:
|
||||||
|
|
||||||
# Disable xformers.
|
# Disable xformers.
|
||||||
|
|||||||
Loading…
Reference in New Issue
Block a user