Skip to content

Commit 3ae082c

Browse files
dongbo910220yewentao256DarkLight1337
authored
[Chore] Separate out optional dependency checks from vllm.utils (#27207)
Signed-off-by: dongbo910220 <[email protected]> Signed-off-by: dongbo910220 <[email protected]> Co-authored-by: Wentao Ye <[email protected]> Co-authored-by: Cyrus Leung <[email protected]>
1 parent 49c00fe commit 3ae082c

File tree

22 files changed

+60
-64
lines changed

22 files changed

+60
-64
lines changed

tests/kernels/attention/test_deepgemm_attention.py

Lines changed: 2 additions & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -6,7 +6,7 @@
66
import torch
77

88
from vllm.platforms import current_platform
9-
from vllm.utils import cdiv, has_deep_gemm
9+
from vllm.utils import cdiv
1010
from vllm.utils.deep_gemm import (
1111
_ceil_to_ue8m0,
1212
calc_diff,
@@ -15,6 +15,7 @@
1515
get_num_sms,
1616
get_paged_mqa_logits_metadata,
1717
)
18+
from vllm.utils.import_utils import has_deep_gemm
1819

1920

2021
def kv_cache_cast_to_fp8(x: torch.Tensor) -> torch.Tensor:

tests/kernels/moe/modular_kernel_tools/common.py

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -23,7 +23,7 @@
2323
FusedMoEQuantConfig,
2424
)
2525
from vllm.model_executor.layers.fused_moe.fused_moe import fused_topk
26-
from vllm.utils import has_deep_ep, has_deep_gemm, has_pplx
26+
from vllm.utils.import_utils import has_deep_ep, has_deep_gemm, has_pplx
2727

2828
from .mk_objects import (
2929
TestMoEQuantConfig,

tests/kernels/moe/modular_kernel_tools/mk_objects.py

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -35,9 +35,9 @@
3535
cutlass_fp8_supported,
3636
)
3737
from vllm.platforms import current_platform
38-
from vllm.utils import has_deep_ep, has_deep_gemm, has_pplx
3938
from vllm.utils.deep_gemm import is_deep_gemm_supported
4039
from vllm.utils.flashinfer import has_flashinfer_cutlass_fused_moe
40+
from vllm.utils.import_utils import has_deep_ep, has_deep_gemm, has_pplx
4141

4242

4343
@dataclass

tests/kernels/moe/parallel_utils.py

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -15,7 +15,7 @@
1515
from torch.multiprocessing import spawn # pyright: ignore[reportPrivateImportUsage]
1616
from typing_extensions import ParamSpec
1717

18-
from vllm.utils import has_deep_ep
18+
from vllm.utils.import_utils import has_deep_ep
1919
from vllm.utils.network_utils import get_open_port
2020

2121
if has_deep_ep():

tests/kernels/moe/test_block_fp8.py

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -21,11 +21,11 @@
2121
modular_triton_fused_moe,
2222
)
2323
from vllm.platforms import current_platform
24-
from vllm.utils import has_deep_gemm
2524
from vllm.utils.deep_gemm import (
2625
get_mk_alignment_for_contiguous_layout,
2726
is_deep_gemm_e8m0_used,
2827
)
28+
from vllm.utils.import_utils import has_deep_gemm
2929

3030
dg_available = has_deep_gemm()
3131

tests/kernels/moe/test_deepep_deepgemm_moe.py

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -21,8 +21,8 @@
2121
from vllm.model_executor.layers.fused_moe.fused_moe import fused_experts
2222
from vllm.model_executor.layers.fused_moe.modular_kernel import FusedMoEModularKernel
2323
from vllm.platforms import current_platform
24-
from vllm.utils import has_deep_ep, has_deep_gemm
2524
from vllm.utils.deep_gemm import is_deep_gemm_e8m0_used, is_deep_gemm_supported
25+
from vllm.utils.import_utils import has_deep_ep, has_deep_gemm
2626

2727
from ...utils import multi_gpu_test
2828
from .parallel_utils import ProcessGroupInfo, parallel_launch

tests/kernels/moe/test_deepep_moe.py

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -21,7 +21,7 @@
2121
per_token_group_quant_fp8,
2222
)
2323
from vllm.platforms import current_platform
24-
from vllm.utils import has_deep_ep
24+
from vllm.utils.import_utils import has_deep_ep
2525

2626
from ...utils import multi_gpu_test
2727
from .parallel_utils import ProcessGroupInfo, parallel_launch

tests/kernels/moe/test_gpt_oss_triton_kernels.py

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -6,7 +6,7 @@
66
import torch
77
import torch.nn.functional as F
88

9-
from vllm.utils import has_triton_kernels
9+
from vllm.utils.import_utils import has_triton_kernels
1010

1111
if not has_triton_kernels():
1212
pytest.skip(

tests/kernels/moe/test_modular_kernel_combinations.py

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -13,8 +13,8 @@
1313
import vllm.model_executor.layers.fused_moe.modular_kernel as mk
1414
from vllm.config import VllmConfig, set_current_vllm_config
1515
from vllm.platforms import current_platform
16-
from vllm.utils import has_deep_ep, has_deep_gemm, has_pplx
1716
from vllm.utils.flashinfer import has_flashinfer_cutlass_fused_moe
17+
from vllm.utils.import_utils import has_deep_ep, has_deep_gemm, has_pplx
1818
from vllm.utils.torch_utils import cuda_device_count_stateless
1919

2020
from .modular_kernel_tools.common import (

tests/kernels/quantization/test_block_fp8.py

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -18,12 +18,12 @@
1818
w8a8_triton_block_scaled_mm,
1919
)
2020
from vllm.platforms import current_platform
21-
from vllm.utils import has_deep_gemm
2221
from vllm.utils.deep_gemm import (
2322
fp8_gemm_nt,
2423
get_col_major_tma_aligned_tensor,
2524
per_block_cast_to_fp8,
2625
)
26+
from vllm.utils.import_utils import has_deep_gemm
2727

2828
if current_platform.get_device_capability() < (9, 0):
2929
pytest.skip("FP8 Triton requires CUDA 9.0 or higher", allow_module_level=True)

0 commit comments

Comments
 (0)