diff --git a/.buildkite/test-pipeline.yaml b/.buildkite/test-pipeline.yaml index cc1223d4c465..4455c2a5dbfb 100644 --- a/.buildkite/test-pipeline.yaml +++ b/.buildkite/test-pipeline.yaml @@ -535,8 +535,6 @@ steps: - vllm/ - tests/models/language commands: - # Install causal-conv1d for plamo2 models here, as it is not compatible with pip-compile. - - pip install 'git+https://github.com/Dao-AILab/causal-conv1d@v1.5.0.post8' - pip freeze | grep -E 'torch' - pytest -v -s models/language -m core_model @@ -547,8 +545,6 @@ steps: - vllm/ - tests/models/language/generation commands: - # Install causal-conv1d for plamo2 models here, as it is not compatible with pip-compile. - - pip install 'git+https://github.com/Dao-AILab/causal-conv1d@v1.5.0.post8' - pytest -v -s models/language/generation -m hybrid_model - label: Language Models Test (Extended Generation) # 1hr20min @@ -558,8 +554,6 @@ steps: - vllm/ - tests/models/language/generation commands: - # Install causal-conv1d for plamo2 models here, as it is not compatible with pip-compile. - - pip install 'git+https://github.com/Dao-AILab/causal-conv1d@v1.5.0.post8' - pytest -v -s models/language/generation -m '(not core_model) and (not hybrid_model)' - label: Language Models Test (Extended Pooling) # 36min diff --git a/docker/Dockerfile.cpu b/docker/Dockerfile.cpu index 1a0981f8ea6d..da9a37a56bc1 100644 --- a/docker/Dockerfile.cpu +++ b/docker/Dockerfile.cpu @@ -113,6 +113,7 @@ WORKDIR /workspace/vllm RUN --mount=type=bind,src=requirements/test.in,target=requirements/test.in \ cp requirements/test.in requirements/cpu-test.in && \ + sed -i '/causal_conv1d/d' requirements/cpu-test.in && \ sed -i '/mamba_ssm/d' requirements/cpu-test.in && \ sed -i 's/^torch==.*/torch==2.6.0/g' requirements/cpu-test.in && \ sed -i 's/torchaudio.*/torchaudio/g' requirements/cpu-test.in && \ diff --git a/docker/Dockerfile.rocm b/docker/Dockerfile.rocm index 4f40f32a39f2..7ac4855ba632 100644 --- a/docker/Dockerfile.rocm +++ b/docker/Dockerfile.rocm @@ -71,7 +71,7 @@ COPY --from=build_vllm ${COMMON_WORKDIR}/vllm /vllm-workspace RUN cd /vllm-workspace \ && rm -rf vllm \ && python3 -m pip install -e tests/vllm_test_utils \ - && python3 -m pip install lm-eval[api]==0.4.4 \ + && python3 -m pip install git+https://github.com/EleutherAI/lm-evaluation-harness.git@bc811365ef3917a15da6aac9b7feafb4ba0fb42f#egg=lm-eval[api] \ && python3 -m pip install pytest-shard # ----------------------- diff --git a/docs/contributing/ci/update_pytorch_version.md b/docs/contributing/ci/update_pytorch_version.md index 3a6026d450a6..5006944005f8 100644 --- a/docs/contributing/ci/update_pytorch_version.md +++ b/docs/contributing/ci/update_pytorch_version.md @@ -141,7 +141,7 @@ uv pip install --system \ ### causal-conv1d ```bash -uv pip install 'git+https://github.com/Dao-AILab/causal-conv1d@v1.5.0.post8' +uv pip install 'git+https://github.com/Dao-AILab/causal-conv1d@v1.5.2' ``` ## Update all the different vLLM platforms diff --git a/requirements/test.in b/requirements/test.in index 9ecaaae92727..effb30ad02b7 100644 --- a/requirements/test.in +++ b/requirements/test.in @@ -26,6 +26,7 @@ torch==2.7.1 torchaudio==2.7.1 torchvision==0.22.1 transformers_stream_generator # required for qwen-vl test +causal_conv1d==1.5.2 # required for plamo2 test mamba_ssm==2.2.5 # required for plamo2 test matplotlib # required for qwen-vl test mistral_common[image,audio] >= 1.8.2 # required for voxtral test @@ -33,7 +34,7 @@ num2words # required for smolvlm test open_clip_torch==2.32.0 # Required for nemotron_vl test opencv-python-headless >= 4.11.0 # required for video test datamodel_code_generator # required for minicpm3 test -lm-eval[api]==0.4.8 # required for model evaluation test +lm-eval[api] @ git+https://github.com/EleutherAI/lm-evaluation-harness.git@bc811365ef3917a15da6aac9b7feafb4ba0fb42f # required for model evaluation test mteb[bm25s]>=1.38.11, <2 # required for mteb test transformers==4.53.2 tokenizers==0.21.1 diff --git a/requirements/test.txt b/requirements/test.txt index 691420df87c4..05e7f4dbe1f5 100644 --- a/requirements/test.txt +++ b/requirements/test.txt @@ -85,6 +85,8 @@ cachetools==5.5.2 # via # google-auth # mlflow-skinny +causal-conv1d==1.5.2 + # via -r requirements/test.in certifi==2024.8.30 # via # fiona @@ -410,7 +412,7 @@ lightning-utilities==0.14.3 # torchmetrics llvmlite==0.44.0 # via numba -lm-eval==0.4.8 +lm-eval @ git+https://github.com/EleutherAI/lm-evaluation-harness.git@bc811365ef3917a15da6aac9b7feafb4ba0fb42f # via -r requirements/test.in lxml==5.3.0 # via @@ -477,7 +479,9 @@ networkx==3.2.1 # scikit-image # torch ninja==1.11.1.3 - # via mamba-ssm + # via + # causal-conv1d + # mamba-ssm nltk==3.9.1 # via rouge-score num2words==0.5.14 @@ -618,6 +622,7 @@ packaging==24.2 # via # accelerate # black + # causal-conv1d # datamodel-code-generator # datasets # evaluate @@ -1079,6 +1084,7 @@ torch==2.7.1+cu128 # -r requirements/test.in # accelerate # bitsandbytes + # causal-conv1d # efficientnet-pytorch # encodec # fastsafetensors diff --git a/tests/models/language/generation/test_hybrid.py b/tests/models/language/generation/test_hybrid.py index 2238924c1b50..ac54f16f5eaf 100644 --- a/tests/models/language/generation/test_hybrid.py +++ b/tests/models/language/generation/test_hybrid.py @@ -25,9 +25,6 @@ HYBRID_MODELS = [ "ai21labs/Jamba-tiny-dev", - # NOTE: Running Plamo2 in transformers implementation requires to install - # causal-conv1d package, which is not listed as a test dependency as it's - # not compatible with pip-compile. "pfnet/plamo-2-1b", "Zyphra/Zamba2-1.2B-instruct", "hmellor/tiny-random-BambaForCausalLM",