-
Notifications
You must be signed in to change notification settings - Fork 72
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
[CI] Add new runner and enable QwQ multinpu test #417
base: main
Are you sure you want to change the base?
Conversation
e8aa80c
to
dd0a9c2
Compare
1599f82
to
ebdad10
Compare
tensor_parallel_size=4, | ||
distributed_executor_backend=distributed_executor_backend, | ||
) as vllm_model: | ||
vllm_model.generate_greedy(example_prompts, max_tokens) |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
I remember last offline discussion, we'd also want to add transformer as compare, does it works (stable works)?
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
will finish at next pull request
Signed-off-by: wangli <[email protected]>
Signed-off-by: wangli <[email protected]>
Signed-off-by: wangli <[email protected]>
Signed-off-by: wangli <[email protected]>
Co-authored-by: Yikun Jiang <[email protected]> Signed-off-by: wangli <[email protected]>
Signed-off-by: wangli <[email protected]>
Signed-off-by: wangli <[email protected]>
Signed-off-by: wangli <[email protected]>
Signed-off-by: wangli <[email protected]>
Signed-off-by: wangli <[email protected]>
Signed-off-by: wangli <[email protected]>
Signed-off-by: wangli <[email protected]>
Signed-off-by: wangli <[email protected]>
Signed-off-by: wangli <[email protected]>
Signed-off-by: wangli <[email protected]>
fi | ||
|
||
pip install /root/.cache/pta/torch_npu-2.5.1.dev20250320-cp310-cp310-manylinux_2_17_aarch64.manylinux2014_aarch64.whl | ||
- name: Run vllm-project/vllm-ascend test |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
V1 engine should be tested as well.
What this PR does / why we need it?
Does this PR introduce any user-facing change?
No
How was this patch tested?
CI passed