mirror of
https://github.com/Comfy-Org/ComfyUI-Manager.git
synced 2025-12-24 13:50:49 +08:00
update DB
This commit is contained in:
parent
c849af712c
commit
36ecc15cf0
@ -344,7 +344,9 @@
|
||||
"Add Folder",
|
||||
"Add Folder Advanced",
|
||||
"Create Project Root",
|
||||
"ShowPath"
|
||||
"Join Variables",
|
||||
"Show Path",
|
||||
"Show String"
|
||||
],
|
||||
{
|
||||
"title_aux": "ComfyUI-Path-Helper"
|
||||
@ -495,6 +497,7 @@
|
||||
"PrimereEmbeddingKeywordMerger",
|
||||
"PrimereHypernetwork",
|
||||
"PrimereImageSegments",
|
||||
"PrimereKSampler",
|
||||
"PrimereLCMSelector",
|
||||
"PrimereLORA",
|
||||
"PrimereLYCORIS",
|
||||
@ -507,6 +510,7 @@
|
||||
"PrimereMetaRead",
|
||||
"PrimereMetaSave",
|
||||
"PrimereMidjourneyStyles",
|
||||
"PrimereModelConceptSelector",
|
||||
"PrimereModelKeyword",
|
||||
"PrimereNetworkTagLoader",
|
||||
"PrimerePrompt",
|
||||
@ -515,6 +519,7 @@
|
||||
"PrimereResolution",
|
||||
"PrimereResolutionMultiplier",
|
||||
"PrimereSamplers",
|
||||
"PrimereSamplersSteps",
|
||||
"PrimereSeed",
|
||||
"PrimereStepsCfg",
|
||||
"PrimereStyleLoader",
|
||||
@ -5260,15 +5265,21 @@
|
||||
"https://github.com/gokayfem/ComfyUI_VLM_nodes": [
|
||||
[
|
||||
"Joytag",
|
||||
"JsonToText",
|
||||
"KeywordExtraction",
|
||||
"LLMLoader",
|
||||
"LLMSamplerAdvanced",
|
||||
"LLMPromptGenerator",
|
||||
"LLMSampler",
|
||||
"LLava Loader Simple",
|
||||
"LLavaPromptGenerator",
|
||||
"LLavaSamplerAdvanced",
|
||||
"LLavaSamplerSimple",
|
||||
"LlavaClipLoader",
|
||||
"MoonDream",
|
||||
"PromptGenerate",
|
||||
"SimpleText"
|
||||
"PromptGenerateAPI",
|
||||
"SimpleText",
|
||||
"Suggester",
|
||||
"ViewText"
|
||||
],
|
||||
{
|
||||
"title_aux": "VLM_nodes"
|
||||
@ -5680,6 +5691,7 @@
|
||||
"ReverseImageBatch",
|
||||
"RoundMask",
|
||||
"SaveImageWithAlpha",
|
||||
"ScaleBatchPromptSchedule",
|
||||
"SomethingToString",
|
||||
"SoundReactive",
|
||||
"SplitBboxes",
|
||||
|
||||
@ -1883,5 +1883,5 @@
|
||||
"filename": "diffusion_pytorch_model.safetensors",
|
||||
"url": "https://huggingface.co/InstantX/InstantID/resolve/main/ControlNetModel/diffusion_pytorch_model.safetensors"
|
||||
}
|
||||
]
|
||||
]
|
||||
}
|
||||
|
||||
@ -344,7 +344,9 @@
|
||||
"Add Folder",
|
||||
"Add Folder Advanced",
|
||||
"Create Project Root",
|
||||
"ShowPath"
|
||||
"Join Variables",
|
||||
"Show Path",
|
||||
"Show String"
|
||||
],
|
||||
{
|
||||
"title_aux": "ComfyUI-Path-Helper"
|
||||
@ -495,6 +497,7 @@
|
||||
"PrimereEmbeddingKeywordMerger",
|
||||
"PrimereHypernetwork",
|
||||
"PrimereImageSegments",
|
||||
"PrimereKSampler",
|
||||
"PrimereLCMSelector",
|
||||
"PrimereLORA",
|
||||
"PrimereLYCORIS",
|
||||
@ -507,6 +510,7 @@
|
||||
"PrimereMetaRead",
|
||||
"PrimereMetaSave",
|
||||
"PrimereMidjourneyStyles",
|
||||
"PrimereModelConceptSelector",
|
||||
"PrimereModelKeyword",
|
||||
"PrimereNetworkTagLoader",
|
||||
"PrimerePrompt",
|
||||
@ -515,6 +519,7 @@
|
||||
"PrimereResolution",
|
||||
"PrimereResolutionMultiplier",
|
||||
"PrimereSamplers",
|
||||
"PrimereSamplersSteps",
|
||||
"PrimereSeed",
|
||||
"PrimereStepsCfg",
|
||||
"PrimereStyleLoader",
|
||||
@ -5260,15 +5265,21 @@
|
||||
"https://github.com/gokayfem/ComfyUI_VLM_nodes": [
|
||||
[
|
||||
"Joytag",
|
||||
"JsonToText",
|
||||
"KeywordExtraction",
|
||||
"LLMLoader",
|
||||
"LLMSamplerAdvanced",
|
||||
"LLMPromptGenerator",
|
||||
"LLMSampler",
|
||||
"LLava Loader Simple",
|
||||
"LLavaPromptGenerator",
|
||||
"LLavaSamplerAdvanced",
|
||||
"LLavaSamplerSimple",
|
||||
"LlavaClipLoader",
|
||||
"MoonDream",
|
||||
"PromptGenerate",
|
||||
"SimpleText"
|
||||
"PromptGenerateAPI",
|
||||
"SimpleText",
|
||||
"Suggester",
|
||||
"ViewText"
|
||||
],
|
||||
{
|
||||
"title_aux": "VLM_nodes"
|
||||
@ -5680,6 +5691,7 @@
|
||||
"ReverseImageBatch",
|
||||
"RoundMask",
|
||||
"SaveImageWithAlpha",
|
||||
"ScaleBatchPromptSchedule",
|
||||
"SomethingToString",
|
||||
"SoundReactive",
|
||||
"SplitBboxes",
|
||||
|
||||
@ -1,5 +1,56 @@
|
||||
{
|
||||
"models": [
|
||||
{
|
||||
"name": "1k3d68.onnx",
|
||||
"type": "insightface",
|
||||
"base": "inswapper",
|
||||
"save_path": "insightface/models/antelopev2",
|
||||
"description": "Antelopev2 1k3d68.onnx model for InstantId. (InstantId needs all Antelopev2 models)",
|
||||
"reference": "https://github.com/cubiq/ComfyUI_InstantID#installation",
|
||||
"filename": "1k3d68.onnx",
|
||||
"url": "https://huggingface.co/MonsterMMORPG/tools/resolve/main/1k3d68.onnx"
|
||||
},
|
||||
{
|
||||
"name": "2d106det.onnx",
|
||||
"type": "insightface",
|
||||
"base": "inswapper",
|
||||
"save_path": "insightface/models/antelopev2",
|
||||
"description": "Antelopev2 2d106det.onnx model for InstantId. (InstantId needs all Antelopev2 models)",
|
||||
"reference": "https://github.com/cubiq/ComfyUI_InstantID#installation",
|
||||
"filename": "2d106det.onnx",
|
||||
"url": "https://huggingface.co/MonsterMMORPG/tools/resolve/main/2d106det.onnx"
|
||||
},
|
||||
{
|
||||
"name": "genderage.onnx",
|
||||
"type": "insightface",
|
||||
"base": "inswapper",
|
||||
"save_path": "insightface/models/antelopev2",
|
||||
"description": "Antelopev2 genderage.onnx model for InstantId. (InstantId needs all Antelopev2 models)",
|
||||
"reference": "https://github.com/cubiq/ComfyUI_InstantID#installation",
|
||||
"filename": "genderage.onnx",
|
||||
"url": "https://huggingface.co/MonsterMMORPG/tools/resolve/main/genderage.onnx"
|
||||
},
|
||||
{
|
||||
"name": "glintr100.onnx",
|
||||
"type": "insightface",
|
||||
"base": "inswapper",
|
||||
"save_path": "insightface/models/antelopev2",
|
||||
"description": "Antelopev2 glintr100.onnx model for InstantId. (InstantId needs all Antelopev2 models)",
|
||||
"reference": "https://github.com/cubiq/ComfyUI_InstantID#installation",
|
||||
"filename": "glintr100.onnx",
|
||||
"url": "https://huggingface.co/MonsterMMORPG/tools/resolve/main/glintr100.onnx"
|
||||
},
|
||||
{
|
||||
"name": "scrfd_10g_bnkps.onnx",
|
||||
"type": "insightface",
|
||||
"base": "inswapper",
|
||||
"save_path": "insightface/models/antelopev2",
|
||||
"description": "Antelopev2 scrfd_10g_bnkps.onnx model for InstantId. (InstantId needs all Antelopev2 models)",
|
||||
"reference": "https://github.com/cubiq/ComfyUI_InstantID#installation",
|
||||
"filename": "scrfd_10g_bnkps.onnx",
|
||||
"url": "https://huggingface.co/MonsterMMORPG/tools/resolve/main/scrfd_10g_bnkps.onnx"
|
||||
},
|
||||
|
||||
{
|
||||
"name": "photomaker-v1.bin",
|
||||
"type": "photomaker",
|
||||
@ -652,131 +703,6 @@
|
||||
"reference": "https://huggingface.co/manshoety/AD_Stabilized_Motion",
|
||||
"filename": "mm-Stabilized_mid.pth",
|
||||
"url": "https://huggingface.co/manshoety/AD_Stabilized_Motion/resolve/main/mm-Stabilized_mid.pth"
|
||||
},
|
||||
|
||||
{
|
||||
"name": "GFPGANv1.4.pth",
|
||||
"type": "GFPGAN",
|
||||
"base": "GFPGAN",
|
||||
"save_path": "facerestore_models",
|
||||
"description": "Face Restoration Models. Download the model required for using the 'Facerestore CF (Code Former)' custom node.",
|
||||
"reference": "https://github.com/TencentARC/GFPGAN/releases",
|
||||
"filename": "GFPGANv1.4.pth",
|
||||
"url": "https://github.com/TencentARC/GFPGAN/releases/download/v1.3.4/GFPGANv1.4.pth"
|
||||
},
|
||||
{
|
||||
"name": "codeformer.pth",
|
||||
"type": "CodeFormer",
|
||||
"base": "CodeFormer",
|
||||
"save_path": "facerestore_models",
|
||||
"description": "Face Restoration Models. Download the model required for using the 'Facerestore CF (Code Former)' custom node.",
|
||||
"reference": "https://github.com/sczhou/CodeFormer/releases",
|
||||
"filename": "codeformer.pth",
|
||||
"url": "https://github.com/sczhou/CodeFormer/releases/download/v0.1.0/codeformer.pth"
|
||||
},
|
||||
{
|
||||
"name": "detection_Resnet50_Final.pth",
|
||||
"type": "facexlib",
|
||||
"base": "facexlib",
|
||||
"save_path": "facerestore_models",
|
||||
"description": "Face Detection Models. Download the model required for using the 'Facerestore CF (Code Former)' custom node.",
|
||||
"reference": "https://github.com/xinntao/facexlib",
|
||||
"filename": "detection_Resnet50_Final.pth",
|
||||
"url": "https://github.com/xinntao/facexlib/releases/download/v0.1.0/detection_Resnet50_Final.pth"
|
||||
},
|
||||
{
|
||||
"name": "detection_mobilenet0.25_Final.pth",
|
||||
"type": "facexlib",
|
||||
"base": "facexlib",
|
||||
"save_path": "facerestore_models",
|
||||
"description": "Face Detection Models. Download the model required for using the 'Facerestore CF (Code Former)' custom node.",
|
||||
"reference": "https://github.com/xinntao/facexlib",
|
||||
"filename": "detection_mobilenet0.25_Final.pth",
|
||||
"url": "https://github.com/xinntao/facexlib/releases/download/v0.1.0/detection_mobilenet0.25_Final.pth"
|
||||
},
|
||||
{
|
||||
"name": "yolov5l-face.pth",
|
||||
"type": "facexlib",
|
||||
"base": "facexlib",
|
||||
"save_path": "facedetection",
|
||||
"description": "Face Detection Models. Download the model required for using the 'Facerestore CF (Code Former)' custom node.",
|
||||
"reference": "https://github.com/xinntao/facexlib",
|
||||
"filename": "yolov5l-face.pth",
|
||||
"url": "https://github.com/sczhou/CodeFormer/releases/download/v0.1.0/yolov5l-face.pth"
|
||||
},
|
||||
{
|
||||
"name": "yolov5n-face.pth",
|
||||
"type": "facexlib",
|
||||
"base": "facexlib",
|
||||
"save_path": "facedetection",
|
||||
"description": "Face Detection Models. Download the model required for using the 'Facerestore CF (Code Former)' custom node.",
|
||||
"reference": "https://github.com/xinntao/facexlib",
|
||||
"filename": "yolov5n-face.pth",
|
||||
"url": "https://github.com/sczhou/CodeFormer/releases/download/v0.1.0/yolov5n-face.pth"
|
||||
},
|
||||
|
||||
{
|
||||
"name": "diffusers/stable-diffusion-xl-1.0-inpainting-0.1 (UNET/fp16)",
|
||||
"type": "unet",
|
||||
"base": "SDXL",
|
||||
"save_path": "unet/xl-inpaint-0.1",
|
||||
"description": "[5.14GB] Stable Diffusion XL inpainting model 0.1. You need UNETLoader instead of CheckpointLoader.",
|
||||
"reference": "https://huggingface.co/diffusers/stable-diffusion-xl-1.0-inpainting-0.1",
|
||||
"filename": "diffusion_pytorch_model.fp16.safetensors",
|
||||
"url": "https://huggingface.co/diffusers/stable-diffusion-xl-1.0-inpainting-0.1/resolve/main/unet/diffusion_pytorch_model.fp16.safetensors"
|
||||
},
|
||||
{
|
||||
"name": "diffusers/stable-diffusion-xl-1.0-inpainting-0.1 (UNET)",
|
||||
"type": "unet",
|
||||
"base": "SDXL",
|
||||
"save_path": "unet/xl-inpaint-0.1",
|
||||
"description": "[10.3GB] Stable Diffusion XL inpainting model 0.1. You need UNETLoader instead of CheckpointLoader.",
|
||||
"reference": "https://huggingface.co/diffusers/stable-diffusion-xl-1.0-inpainting-0.1",
|
||||
"filename": "diffusion_pytorch_model.safetensors",
|
||||
"url": "https://huggingface.co/diffusers/stable-diffusion-xl-1.0-inpainting-0.1/resolve/main/unet/diffusion_pytorch_model.safetensors"
|
||||
},
|
||||
|
||||
{
|
||||
"name": "Inswapper (face swap)",
|
||||
"type": "insightface",
|
||||
"base" : "inswapper",
|
||||
"save_path": "insightface",
|
||||
"description": "Checkpoint of the insightface swapper model (used by Comfy-Roop and comfy_mtb)",
|
||||
"reference": "https://huggingface.co/deepinsight/inswapper/",
|
||||
"filename": "inswapper_128.onnx",
|
||||
"url": "https://huggingface.co/deepinsight/inswapper/resolve/main/inswapper_128.onnx"
|
||||
},
|
||||
|
||||
{
|
||||
"name": "CLIPVision model (stabilityai/clip_vision_g)",
|
||||
"type": "clip_vision",
|
||||
"base": "vit-g",
|
||||
"save_path": "clip_vision",
|
||||
"description": "[3.69GB] clip_g vision model",
|
||||
"reference": "https://huggingface.co/stabilityai/control-lora",
|
||||
"filename": "clip_vision_g.safetensors",
|
||||
"url": "https://huggingface.co/stabilityai/control-lora/resolve/main/revision/clip_vision_g.safetensors"
|
||||
},
|
||||
|
||||
{
|
||||
"name": "CLIPVision model (IP-Adapter) CLIP-ViT-H-14-laion2B-s32B-b79K",
|
||||
"type": "clip_vision",
|
||||
"base": "ViT-H",
|
||||
"save_path": "clip_vision",
|
||||
"description": "[2.5GB] CLIPVision model (needed for IP-Adapter)",
|
||||
"reference": "https://huggingface.co/h94/IP-Adapter",
|
||||
"filename": "CLIP-ViT-H-14-laion2B-s32B-b79K.safetensors",
|
||||
"url": "https://huggingface.co/h94/IP-Adapter/resolve/main/models/image_encoder/model.safetensors"
|
||||
},
|
||||
{
|
||||
"name": "CLIPVision model (IP-Adapter) CLIP-ViT-bigG-14-laion2B-39B-b160k",
|
||||
"type": "clip_vision",
|
||||
"base": "ViT-G",
|
||||
"save_path": "clip_vision",
|
||||
"description": "[3.69GB] CLIPVision model (needed for IP-Adapter)",
|
||||
"reference": "https://huggingface.co/h94/IP-Adapter",
|
||||
"filename": "CLIP-ViT-bigG-14-laion2B-39B-b160k.safetensors",
|
||||
"url": "https://huggingface.co/h94/IP-Adapter/resolve/main/sdxl_models/image_encoder/model.safetensors"
|
||||
}
|
||||
]
|
||||
}
|
||||
|
||||
Loading…
Reference in New Issue
Block a user