Skip to content

Conversation

@hmellor
Copy link
Member

@hmellor hmellor commented Oct 27, 2025

@mergify
Copy link

mergify bot commented Oct 27, 2025

Documentation preview: https://vllm--27593.org.readthedocs.build/en/27593/

@mergify mergify bot added documentation Improvements or additions to documentation llama Related to Llama models v1 labels Oct 27, 2025
Copy link
Contributor

@gemini-code-assist gemini-code-assist bot left a comment

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

Code Review

This pull request removes the deprecated use_cudagraph and full_cuda_graph fields from the CompilationConfig and updates the config tests accordingly. The removal of these flags simplifies the configuration and aligns with the deprecation schedule. The test updates ensure the config tests are passing.

Signed-off-by: Harry Mellor <[email protected]>
@ZJY0516
Copy link
Contributor

ZJY0516 commented Oct 28, 2025

CC @fhl2000 @ProExpertProg

Copy link
Contributor

@fhl2000 fhl2000 left a comment

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

LGTM! Thanks for doing the deprecations!

Copy link
Collaborator

@simon-mo simon-mo left a comment

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

stamp

@hmellor hmellor enabled auto-merge (squash) October 28, 2025 20:43
@github-actions github-actions bot added the ready ONLY add when PR is ready to merge/full CI is needed label Oct 28, 2025
vllm_config = VllmConfig(
compilation_config=CompilationConfig(
mode=CompilationMode.VLLM_COMPILE,
use_cudagraph=True,
Copy link
Collaborator

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

Can you make mode explicitly Piecewise here?

# de-duplicate the sizes provided by the config
dedup_sizes = list(set(self.compilation_config.cudagraph_capture_sizes))
cudagraph_capture_sizes = dedup_sizes
# filter out sizes larger than max_cudagraph_capture_size
Copy link
Collaborator

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

Why does this have to be added? Bad merge?

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment

Labels

documentation Improvements or additions to documentation llama Related to Llama models ready ONLY add when PR is ready to merge/full CI is needed v1

Projects

None yet

Development

Successfully merging this pull request may close these issues.

5 participants