mirror of
https://github.com/comfyanonymous/ComfyUI.git
synced 2026-01-25 13:50:15 +08:00
Merge branch 'comfyanonymous:master' into master
This commit is contained in:
commit
1040220970
@ -9,8 +9,14 @@ class AppSettings():
|
|||||||
self.user_manager = user_manager
|
self.user_manager = user_manager
|
||||||
|
|
||||||
def get_settings(self, request):
|
def get_settings(self, request):
|
||||||
file = self.user_manager.get_request_user_filepath(
|
try:
|
||||||
request, "comfy.settings.json")
|
file = self.user_manager.get_request_user_filepath(
|
||||||
|
request,
|
||||||
|
"comfy.settings.json"
|
||||||
|
)
|
||||||
|
except KeyError as e:
|
||||||
|
logging.error("User settings not found.")
|
||||||
|
raise web.HTTPUnauthorized() from e
|
||||||
if os.path.isfile(file):
|
if os.path.isfile(file):
|
||||||
try:
|
try:
|
||||||
with open(file) as f:
|
with open(file) as f:
|
||||||
|
|||||||
@ -79,6 +79,7 @@ fpte_group.add_argument("--fp8_e4m3fn-text-enc", action="store_true", help="Stor
|
|||||||
fpte_group.add_argument("--fp8_e5m2-text-enc", action="store_true", help="Store text encoder weights in fp8 (e5m2 variant).")
|
fpte_group.add_argument("--fp8_e5m2-text-enc", action="store_true", help="Store text encoder weights in fp8 (e5m2 variant).")
|
||||||
fpte_group.add_argument("--fp16-text-enc", action="store_true", help="Store text encoder weights in fp16.")
|
fpte_group.add_argument("--fp16-text-enc", action="store_true", help="Store text encoder weights in fp16.")
|
||||||
fpte_group.add_argument("--fp32-text-enc", action="store_true", help="Store text encoder weights in fp32.")
|
fpte_group.add_argument("--fp32-text-enc", action="store_true", help="Store text encoder weights in fp32.")
|
||||||
|
fpte_group.add_argument("--bf16-text-enc", action="store_true", help="Store text encoder weights in bf16.")
|
||||||
|
|
||||||
parser.add_argument("--force-channels-last", action="store_true", help="Force channels last format when inferencing the models.")
|
parser.add_argument("--force-channels-last", action="store_true", help="Force channels last format when inferencing the models.")
|
||||||
|
|
||||||
|
|||||||
@ -824,6 +824,8 @@ def text_encoder_dtype(device=None):
|
|||||||
return torch.float8_e5m2
|
return torch.float8_e5m2
|
||||||
elif args.fp16_text_enc:
|
elif args.fp16_text_enc:
|
||||||
return torch.float16
|
return torch.float16
|
||||||
|
elif args.bf16_text_enc:
|
||||||
|
return torch.bfloat16
|
||||||
elif args.fp32_text_enc:
|
elif args.fp32_text_enc:
|
||||||
return torch.float32
|
return torch.float32
|
||||||
|
|
||||||
|
|||||||
@ -244,7 +244,7 @@ def save_glb(vertices, faces, filepath, metadata=None):
|
|||||||
|
|
||||||
Parameters:
|
Parameters:
|
||||||
vertices: torch.Tensor of shape (N, 3) - The vertex coordinates
|
vertices: torch.Tensor of shape (N, 3) - The vertex coordinates
|
||||||
faces: torch.Tensor of shape (M, 4) or (M, 3) - The face indices (quad or triangle faces)
|
faces: torch.Tensor of shape (M, 3) - The face indices (triangle faces)
|
||||||
filepath: str - Output filepath (should end with .glb)
|
filepath: str - Output filepath (should end with .glb)
|
||||||
"""
|
"""
|
||||||
|
|
||||||
|
|||||||
Loading…
Reference in New Issue
Block a user