Skip to content

Conversation

@Gongdayao
Copy link

@Gongdayao Gongdayao commented Dec 3, 2025

What this PR does / why we need it?

This PR adds tutorials for the DeepSeeK-R1 series models, including the A2 and A3 series, and provides accuracy validation results.

Does this PR introduce any user-facing change?

How was this patch tested?

Copy link
Contributor

@gemini-code-assist gemini-code-assist bot left a comment

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

Code Review

This pull request adds a new tutorial for deploying the DeepSeek-R1 model. The documentation is well-structured and provides detailed instructions for both single-node and multi-node deployments. However, I found some critical inconsistencies in the model name within the provided commands, which will cause them to fail. Correcting these will significantly improve the user experience of following this tutorial.

Comment on lines +262 to +285
lm_eval \
--model local-completions \
--model_args model=path/DeepSeek-R1-w8a8,base_url=http://<node0_ip>:<port>/v1/completions,tokenized_requests=False,trust_remote_code=True \
--tasks gsm8k \
--output_path ./
Copy link
Contributor

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

high

The model path path/DeepSeek-R1-w8a8 uses w8a8 in lowercase, which is inconsistent with the model name DeepSeek-R1-W8A8 used in the download link and the vllm serve commands. This will cause a 'file not found' error on case-sensitive filesystems.

Please correct this to DeepSeek-R1-W8A8 to ensure the command works as expected. Similar inconsistencies are present on lines 6, 18, and 291 and should also be corrected.

Suggested change
lm_eval \
--model local-completions \
--model_args model=path/DeepSeek-R1-w8a8,base_url=http://<node0_ip>:<port>/v1/completions,tokenized_requests=False,trust_remote_code=True \
--tasks gsm8k \
--output_path ./
lm_eval \
--model local-completions \
--model_args model=path/DeepSeek-R1-W8A8,base_url=http://<node0_ip>:<port>/v1/completions,tokenized_requests=False,trust_remote_code=True \
--tasks gsm8k \
--output_path ./


```shell
export VLLM_USE_MODELSCOPE=true
vllm bench serve --model path/DeepSeek-R1-w8a8 --dataset-name random --random-input 200 --num-prompt 200 --request-rate 1 --save-result --result-dir ./
Copy link
Contributor

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

high

The model path path/DeepSeek-R1-w8a8 is inconsistent with the official model name DeepSeek-R1-W8A8 from the download link. This will cause a 'file not found' error. Please correct it to DeepSeek-R1-W8A8.

Suggested change
vllm bench serve --model path/DeepSeek-R1-w8a8 --dataset-name random --random-input 200 --num-prompt 200 --request-rate 1 --save-result --result-dir ./
vllm bench serve --model path/DeepSeek-R1-W8A8 --dataset-name random --random-input 200 --num-prompt 200 --request-rate 1 --save-result --result-dir ./

@github-actions
Copy link

github-actions bot commented Dec 3, 2025

👋 Hi! Thank you for contributing to the vLLM Ascend project. The following points will speed up your PR merge:‌‌

  • A PR should do only one thing, smaller PRs enable faster reviews.
  • Every PR should include unit tests and end-to-end tests ‌to ensure it works and is not broken by other future PRs.
  • Write the commit message by fulfilling the PR description to help reviewer and future developers understand.

If CI fails, you can run linting and testing checks locally according Contributing and Testing.

@github-actions github-actions bot added the documentation Improvements or additions to documentation label Dec 3, 2025
::::{tab-item} A3 series
:sync: A3

1. Start the docker image on your node, refer to [using docker](../installation.md#set-up-using-docker).
Copy link
Contributor

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

provide docker run command directly, like "#4399"

Comment on lines 107 to 131
::::
::::{tab-item} DeepSeek-R1-w8a8 A2 series
:sync: A2

Run the following scripts on two nodes respectively.

Copy link
Contributor

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

provide docker run command by refering A3 series

@1092626063
Copy link
Contributor

/lgtm

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment

Labels

documentation Improvements or additions to documentation

Projects

None yet

Development

Successfully merging this pull request may close these issues.

2 participants