Skip to content

Commit 84d2a72

Browse files
jiawenliu64facebook-github-bot
authored andcommitted
Fix MXFP8 test due to OSS FBGEMM version mismatch issue (#4749)
Summary: Pull Request resolved: #4749 X-link: facebookresearch/FBGEMM#1771 Currently there is a version mismatch issue in OSS FBGEMM. Move the latest added `to_mxfp8` to `test_mx_grouped_gemm` to release the issue in tests Will add it back when the OSS FBGEMM version mismatch issue is resolved Reviewed By: q10 Differential Revision: D80634475 fbshipit-source-id: c77f0b0226242c7a0235a6738b1acffc88f6b323
1 parent da77802 commit 84d2a72

File tree

1 file changed

+2
-1
lines changed

1 file changed

+2
-1
lines changed

fbgemm_gpu/experimental/gen_ai/test/quantize/quantize_test.py

Lines changed: 2 additions & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -24,7 +24,6 @@
2424
quantize_fp8_block,
2525
quantize_fp8_row,
2626
supports_float8_fnuz,
27-
to_mxfp8,
2827
)
2928

3029
from fbgemm_gpu.experimental.gen_ai.quantize import quantize_int4_preshuffle
@@ -1240,6 +1239,8 @@ def test_mx_grouped_gemm(
12401239
N: int,
12411240
K: int,
12421241
) -> None:
1242+
from fbgemm_gpu.experimental.gemm.triton_gemm.fp8_gemm import to_mxfp8
1243+
12431244
X = torch.randn((G, M, K), dtype=torch.bfloat16, device=self.device) * 0.1
12441245
W = torch.randn((G, N, K), dtype=torch.bfloat16, device=self.device) * 0.01
12451246

0 commit comments

Comments
 (0)