Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

[CI] Add new runner and enable QwQ multinpu test #417

Open
wants to merge 15 commits into
base: main
Choose a base branch
from

Conversation

Potabk
Copy link
Contributor

@Potabk Potabk commented Mar 28, 2025

What this PR does / why we need it?

  • Add a new runner to the continuous integration system and keep the original CI runner until the new runner runs stably
  • Add distributed test cases

Does this PR introduce any user-facing change?

No

How was this patch tested?

CI passed

@Potabk Potabk force-pushed the ci branch 2 times, most recently from e8aa80c to dd0a9c2 Compare March 31, 2025 01:34
@MengqingCao MengqingCao mentioned this pull request Mar 31, 2025
21 tasks
@Potabk Potabk force-pushed the ci branch 4 times, most recently from 1599f82 to ebdad10 Compare April 1, 2025 02:18
@Yikun Yikun changed the title [CI]Add new runner [CI] Add new runner and enable QwQ multinpu test Apr 1, 2025
tensor_parallel_size=4,
distributed_executor_backend=distributed_executor_backend,
) as vllm_model:
vllm_model.generate_greedy(example_prompts, max_tokens)
Copy link
Collaborator

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

I remember last offline discussion, we'd also want to add transformer as compare, does it works (stable works)?

https://github.com/vllm-project/vllm/blob/c7e63aa4d84de4f0b076d2974d30cd1cd34a4191/tests/basic_correctness/test_basic_correctness.py#L152

Copy link
Contributor Author

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

will finish at next pull request

Potabk and others added 14 commits April 1, 2025 15:34
Signed-off-by: wangli <[email protected]>
Signed-off-by: wangli <[email protected]>
Signed-off-by: wangli <[email protected]>
Signed-off-by: wangli <[email protected]>
Co-authored-by: Yikun Jiang <[email protected]>
Signed-off-by: wangli <[email protected]>
Signed-off-by: wangli <[email protected]>
Signed-off-by: wangli <[email protected]>
Signed-off-by: wangli <[email protected]>
Signed-off-by: wangli <[email protected]>
Signed-off-by: wangli <[email protected]>
Signed-off-by: wangli <[email protected]>
Signed-off-by: wangli <[email protected]>
Signed-off-by: wangli <[email protected]>
fi

pip install /root/.cache/pta/torch_npu-2.5.1.dev20250320-cp310-cp310-manylinux_2_17_aarch64.manylinux2014_aarch64.whl
- name: Run vllm-project/vllm-ascend test
Copy link
Collaborator

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

V1 engine should be tested as well.

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Projects
None yet
Development

Successfully merging this pull request may close these issues.

3 participants