mirror of
https://github.com/comfyanonymous/ComfyUI.git
synced 2026-02-16 16:32:34 +08:00
Rename quant dtype parameter
This commit is contained in:
parent
59a2e8c74e
commit
135d3025ea
@ -122,7 +122,7 @@ class TestTensorCoreFP8Layout(unittest.TestCase):
|
|||||||
qdata, layout_params = TensorCoreFP8Layout.quantize(
|
qdata, layout_params = TensorCoreFP8Layout.quantize(
|
||||||
float_tensor,
|
float_tensor,
|
||||||
scale=scale,
|
scale=scale,
|
||||||
fp8_dtype=torch.float8_e4m3fn
|
dtype=torch.float8_e4m3fn
|
||||||
)
|
)
|
||||||
|
|
||||||
self.assertEqual(qdata.dtype, torch.float8_e4m3fn)
|
self.assertEqual(qdata.dtype, torch.float8_e4m3fn)
|
||||||
@ -139,7 +139,7 @@ class TestTensorCoreFP8Layout(unittest.TestCase):
|
|||||||
qdata, layout_params = TensorCoreFP8Layout.quantize(
|
qdata, layout_params = TensorCoreFP8Layout.quantize(
|
||||||
float_tensor,
|
float_tensor,
|
||||||
scale=scale,
|
scale=scale,
|
||||||
fp8_dtype=torch.float8_e4m3fn
|
dtype=torch.float8_e4m3fn
|
||||||
)
|
)
|
||||||
|
|
||||||
dequantized = TensorCoreFP8Layout.dequantize(qdata, **layout_params)
|
dequantized = TensorCoreFP8Layout.dequantize(qdata, **layout_params)
|
||||||
|
|||||||
Loading…
Reference in New Issue
Block a user