diff --git a/custom-node-list.json b/custom-node-list.json index febbb2bd..3d1bb9fe 100644 --- a/custom-node-list.json +++ b/custom-node-list.json @@ -1249,7 +1249,7 @@ "https://github.com/laksjdjf/IPAdapter-ComfyUI" ], "install_type": "git-clone", - "description": "Nodes:Load IPAdapter. This custom nodes provides loader of the IP-Adapter model." + "description": "Nodes:Load IPAdapter. This custom nodes provides loader of the IP-Adapter model.
NOTE: To use this extension node, you need to download the ip-adapter_sd15.bin file and place it in the custom_nodes/IPAdapter-ComfyUI/models directory. Additionally, you need to download the 'Clip vision model' from the 'Install models' menu as well.
" }, { "author": "taabata", diff --git a/model-list.json b/model-list.json index 6d25f1d5..2bfd1adf 100644 --- a/model-list.json +++ b/model-list.json @@ -451,15 +451,25 @@ "url": "https://huggingface.co/TencentARC/T2I-Adapter/resolve/main/models/t2iadapter_style_sd14v1.pth" }, { - "name": "CLIPVision model", + "name": "CLIPVision model (openai/clip-vit-large)", "type": "clip_vision", "base": "SD1.5", "save_path": "default", - "description": "CLIPVision model (needed for styles model)", + "description": "[1.7GB] CLIPVision model (needed for styles model)", "reference": "https://huggingface.co/openai/clip-vit-large-patch14", "filename": "pytorch_model.bin", "url": "https://huggingface.co/openai/clip-vit-large-patch14/resolve/main/pytorch_model.bin" }, + { + "name": "CLIPVision model (IP-Adapter)", + "type": "clip_vision", + "base": "SD1.5", + "save_path": "default", + "description": "[2.5GB] CLIPVision model (needed for IP-Adapter)", + "reference": "https://huggingface.co/h94/IP-Adapter", + "filename": "pytorch_model.bin", + "url": "https://huggingface.co/h94/IP-Adapter/resolve/main/models/image_encoder/pytorch_model.bin" + }, { "name": "SDXL-controlnet: OpenPose (v2)", "type": "controlnet",