Skip to content

Commit af49ffc

Browse files
authored
Merge branch 'vllm-project:main' into Eagle-mulitmodal-support-Qwen2.5vl
2 parents 36cae1f + a353bd0 commit af49ffc

File tree

1 file changed

+1
-12
lines changed

1 file changed

+1
-12
lines changed

tests/entrypoints/openai/test_default_mm_loras.py

Lines changed: 1 addition & 12 deletions
Original file line numberDiff line numberDiff line change
@@ -24,18 +24,7 @@
2424

2525

2626
@pytest.fixture(scope="module")
27-
def monkeypatch_module():
28-
from _pytest.monkeypatch import MonkeyPatch
29-
mpatch = MonkeyPatch()
30-
yield mpatch
31-
mpatch.undo()
32-
33-
34-
@pytest.fixture(scope="module", params=[False, True])
35-
def multimodal_server(request, monkeypatch_module): # noqa: F811
36-
37-
use_v1 = request.param
38-
monkeypatch_module.setenv('VLLM_USE_V1', '1' if use_v1 else '0')
27+
def multimodal_server(): # noqa: F811
3928

4029
args = [
4130
# use half precision for speed and memory savings in CI environment

0 commit comments

Comments
 (0)