Commit Graph

25 Commits

Author SHA1 Message Date
comfyanonymous
df8cc5d9e3 Add hypernetwork example link to readme.
Move hypernetwork loader node to loaders.
2023-04-24 03:08:51 -04:00
comfyanonymous
5366170a23 Support all known hypernetworks. 2023-04-24 02:36:06 -04:00
comfyanonymous
e6771d0986 Implement Linear hypernetworks.
Add a HypernetworkLoader node to use hypernetworks.
2023-04-23 12:35:25 -04:00
comfyanonymous
4df70d0f62 Fix model_management import so it doesn't get executed twice. 2023-04-15 19:04:33 -04:00
comfyanonymous
25d89b66a3 Fix for older python.
from: https://github.com/comfyanonymous/ComfyUI/discussions/476
2023-04-15 10:56:15 -04:00
comfyanonymous
fb39162c40 LatentCompositeMasked: negative x, y don't work. 2023-04-14 00:49:19 -04:00
comfyanonymous
3fe8074417 Refactor: move nodes_mask_convertion nodes to nodes_mask. 2023-04-14 00:21:01 -04:00
comfyanonymous
7f6c42d878 Merge branch 'image-to-mask' of https://github.com/missionfloyd/ComfyUI
# Conflicts:
#	nodes.py
2023-04-14 00:15:48 -04:00
comfyanonymous
eb3fe3df91 Update comfy_extras/nodes_mask.py
Co-authored-by: missionfloyd <missionfloyd@users.noreply.github.com>
2023-04-14 00:12:15 -04:00
missionfloyd
cb25768a09 Move mask conversion to separate file 2023-04-13 03:11:17 -06:00
mligaintart
c80d164c5b Adds masking to Latent Composite, and provides new masking utilities to
allow better compositing.
2023-04-06 15:18:20 -04:00
comfyanonymous
b09e510414 Rename and reorganize post processing nodes. 2023-04-04 22:54:33 -04:00
comfyanonymous
1f0a236864 Convert line endings to unix. 2023-04-04 13:56:13 -04:00
EllangoK
2313227ae6 use common_upcale in blend 2023-04-04 10:57:34 -04:00
EllangoK
aad55ad912 blend supports any size, dither -> quantize 2023-04-03 09:52:04 -04:00
EllangoK
f9f6352506 adds Blend, Blur, Dither, Sharpen nodes 2023-04-02 18:44:27 -04:00
comfyanonymous
b55667284c Add support for unCLIP SD2.x models.
See _for_testing/unclip in the UI for the new nodes.

unCLIPCheckpointLoader is used to load them.

unCLIPConditioning is used to add the image cond and takes as input a
CLIPVisionEncode output which has been moved to the conditioning section.
2023-04-01 23:19:15 -04:00
comfyanonymous
d5bf8038c8 Merge T2IAdapterLoader and ControlNetLoader.
Workflows will be auto updated.
2023-03-17 18:17:59 -04:00
comfyanonymous
ee6cca1d0f Add folder_paths so models can be in multiple paths. 2023-03-17 18:01:11 -04:00
comfyanonymous
a2136acb6c Prevent model_management from being loaded twice. 2023-03-15 15:18:18 -04:00
comfyanonymous
9c4c183708 Put image upscaling nodes in image/upscaling category. 2023-03-11 18:10:36 -05:00
comfyanonymous
77290f7110 Tiled upscaling with the upscale models. 2023-03-11 14:04:13 -05:00
comfyanonymous
b5cde7df94 Take some code from chainner to implement ESRGAN and other upscale models. 2023-03-11 13:09:28 -05:00
comfyanonymous
75f54df180 A tiny bit of reorganizing. 2023-03-06 01:30:17 -05:00
comfyanonymous
663fa6eafd Implement support for t2i style model.
It needs the CLIPVision model so I added CLIPVisionLoader and CLIPVisionEncode.

Put the clip vision model in models/clip_vision
Put the t2i style model in models/style_models

StyleModelLoader to load it, StyleModelApply to apply it
ConditioningAppend to append the conditioning it outputs to a positive one.
2023-03-05 18:39:25 -05:00