Skip to content

Conversation

ganyi1996ppo
Copy link
Collaborator

@ganyi1996ppo ganyi1996ppo commented Jul 31, 2025

What this PR does / why we need it?

We notice that vllm's main branch merged the PR vllm-project/vllm#21072 and vllm-project/vllm#21473 to support ray backend and fix some rebase bug from previous change. Those changes makes the disaggregate pd in vllm ascend breaks in some scenario.

In this PR, we adopt those changes to make sure the llmdatddist_c_mgr_connector works fine on the newest vllm main branch.

Does this PR introduce any user-facing change?

No user face change.

How was this patch tested?

relevant ut will be added to make sure the functionality of those changes.

@ganyi1996ppo ganyi1996ppo marked this pull request as ready for review July 31, 2025 03:12
Copy link

👋 Hi! Thank you for contributing to the vLLM Ascend project. The following points will speed up your PR merge:‌‌

  • A PR should do only one thing, smaller PRs enable faster reviews.
  • Every PR should include unit tests and end-to-end tests ‌to ensure it works and is not broken by other future PRs.
  • Write the commit message by fulfilling the PR description to help reviewer and future developers understand.

If CI fails, you can run linting and testing checks locally according Contributing and Testing.

@Potabk
Copy link
Collaborator

Potabk commented Aug 1, 2025

looks good if CI passed

Signed-off-by: ganyi <[email protected]>
output = new_output

assert isinstance(output, ModelRunnerOutput)
return output if self.is_driver_worker else None
Copy link
Collaborator

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

can this be removed? I suppose this is usefull in spmd case

Copy link
Collaborator Author

@ganyi1996ppo ganyi1996ppo Aug 1, 2025

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

vllm's main branch removed this in https://github.com/vllm-project/vllm/pull/21473/files , now every tp return its own result, so the tp scenario will encounter the None input case with our previous code

Copy link
Collaborator

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

emm, but it may break PP function I guess. @MengqingCao can you take a check as well?

Copy link
Collaborator

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

I think this is is ok with pp, LGTM

Copy link
Collaborator

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

Double checked with contributor for PP and RL case. It's OK to do the change

Copy link

codecov bot commented Aug 1, 2025

Codecov Report

✅ All modified and coverable lines are covered by tests.
✅ Project coverage is 76.67%. Comparing base (72eceff) to head (e61d0eb).
⚠️ Report is 24 commits behind head on main.

Additional details and impacted files
@@            Coverage Diff             @@
##             main    #2122      +/-   ##
==========================================
+ Coverage   74.41%   76.67%   +2.26%     
==========================================
  Files         100      107       +7     
  Lines       11208    11968     +760     
==========================================
+ Hits         8340     9177     +837     
+ Misses       2868     2791      -77     
Flag Coverage Δ
unittests 76.67% <ø> (+2.26%) ⬆️

Flags with carried forward coverage won't be shown. Click here to find out more.

☔ View full report in Codecov by Sentry.
📢 Have feedback on the report? Share it here.

🚀 New features to boost your workflow:
  • ❄️ Test Analytics: Detect flaky tests, report on failures, and find test suite problems.

output = new_output

assert isinstance(output, ModelRunnerOutput)
return output if self.is_driver_worker else None
Copy link
Collaborator

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

emm, but it may break PP function I guess. @MengqingCao can you take a check as well?

@wangxiyuan wangxiyuan merged commit f939381 into vllm-project:main Aug 4, 2025
25 checks passed
hidden_states: torch.Tensor,
num_scheduled_tokens: int,
num_scheduled_tokens_np: np.ndarray,
finished_sending: Optional[set[str]],
Copy link
Collaborator

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

vLLM now use kv_connector_metadata to store finished_sending and finished_receiving, @Potabk will fix it in a new PR

zzhx1 pushed a commit to lidenghui1110/vllm-ascend that referenced this pull request Aug 11, 2025
…nch (vllm-project#2122)

### What this PR does / why we need it?
We notice that vllm's main branch merged the PR
vllm-project/vllm#21072 and
vllm-project/vllm#21473 to support ray backend
and fix some rebase bug from previous change. Those changes makes the
disaggregate pd in vllm ascend breaks in some scenario.

In this PR, we adopt those changes to make sure the
`llmdatddist_c_mgr_connector` works fine on the newest vllm main branch.

### Does this PR introduce _any_ user-facing change?

No user face change.

### How was this patch tested?
relevant ut will be added to make sure the functionality of those
changes.

- vLLM version: v0.10.0
- vLLM main:
vllm-project/vllm@ad57f23

---------

Signed-off-by: ganyi <[email protected]>
zzhx1 pushed a commit to lidenghui1110/vllm-ascend that referenced this pull request Aug 11, 2025
…nch (vllm-project#2122)

### What this PR does / why we need it?
We notice that vllm's main branch merged the PR
vllm-project/vllm#21072 and
vllm-project/vllm#21473 to support ray backend
and fix some rebase bug from previous change. Those changes makes the
disaggregate pd in vllm ascend breaks in some scenario.

In this PR, we adopt those changes to make sure the
`llmdatddist_c_mgr_connector` works fine on the newest vllm main branch.

### Does this PR introduce _any_ user-facing change?

No user face change.

### How was this patch tested?
relevant ut will be added to make sure the functionality of those
changes.

- vLLM version: v0.10.0
- vLLM main:
vllm-project/vllm@ad57f23

---------

Signed-off-by: ganyi <[email protected]>
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

Successfully merging this pull request may close these issues.

4 participants