revert fp16 support for Lumina2

This commit is contained in:
vanDuven 2025-12-11 01:06:26 -05:00
parent a8ea6953ec
commit 5905513e32

View File

@ -1001,7 +1001,7 @@ class Lumina2(supported_models_base.BASE):
unet_extra_config = {}
latent_format = latent_formats.Flux
supported_inference_dtypes = [torch.bfloat16, torch.float16, torch.float32]
supported_inference_dtypes = [torch.bfloat16, torch.float32]
vae_key_prefix = ["vae."]
text_encoder_key_prefix = ["text_encoders."]