-
Notifications
You must be signed in to change notification settings - Fork 624
[doc] Add Qwen2.5 tutorials #4636
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
base: main
Are you sure you want to change the base?
Conversation
Model module is useless now. Let't remove it totally. - vLLM version: v0.11.2 Signed-off-by: yangshihao6 <[email protected]>
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
Code Review
This pull request adds a new tutorial for deploying the Qwen2.5-7B-Instruct model. The tutorial is comprehensive but appears to be an internal document that has been published without being adapted for a public audience. It contains several internal links, references to internal Docker images, and network-specific instructions that make it impossible for external users to follow. My review focuses on highlighting these issues to ensure the documentation is accessible and reproducible for the open-source community.
docs/source/tutorials/Qwen2.5-7b.md
Outdated
| -v /usr/local/Ascend/driver:/usr/local/Ascend/driver \ | ||
| -v /home/:/home \ | ||
| -w /home/<name> \ | ||
| mindie:dev-2.2.RC1.B070-800I-A2-py312-ubuntu22.03-x86_64 \ |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
The Docker image mindie:dev-2.2.RC1.B070-800I-A2-py312-ubuntu22.03-x86_64 appears to be an internal image, which is not available to the public. For this tutorial to be reproducible by the open-source community, please use a publicly available base image and provide instructions to build upon it if necessary.
docs/source/tutorials/Qwen2.5.md
Outdated
| :::{note} | ||
| The server may be in a restricted network zone (Yellow Zone) and require a Green Zone proxy tool for internet access. Download the proxy tool from the internal repository, run `PortMapping.exe` to obtain the proxy IP, and update `ip_addr` in `portproxy_remote.sh` before executing the script. | ||
| ::: |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
The note about "Yellow Zone", "Green Zone proxy tool", and internal repositories is specific to a corporate network environment. This information is not relevant or helpful for external users and makes the documentation confusing. Please remove these internal-specific instructions to make the guide generally applicable.
| Qwen2.5-7B-Instruct supports single-node single-card deployment on the 910B4 platform. Follow these steps to start the inference service: | ||
|
|
||
| 1. Prepare model weights: Ensure the downloaded model weights are stored in the `./Qwen2.5-7B-Instruct/` directory. | ||
| 2. Download the gsm8k dataset (for evaluation): [gsm8k.zip](https://vision-file-storage/api/file/download/attachment-v2/WIKI202511118986704/32978033/20251111T144846Z_9658c67a0fb349f9be081ab9ab9fd2bc.zip?attachment_id=32978033) |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
docs/source/tutorials/Qwen2.5.md
Outdated
|
|
||
| #### Evaluation Results | ||
| Results and logs are saved to `benchmark/outputs/default/`. A sample accuracy report is shown below: | ||
|  |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
The image for "Accuracy Evaluation Result" is hosted on an internal wiki.huawei.com server and is inaccessible to the public. To make the documentation useful for everyone, please embed the image directly into the repository and use a relative path. This also applies to the "Performance Evaluation Result" image on line 307.
|
👋 Hi! Thank you for contributing to the vLLM Ascend project. The following points will speed up your PR merge:
If CI fails, you can run linting and testing checks locally according Contributing and Testing. |
Model module is useless now. Let't remove it totally. - vLLM version: v0.11.2 Signed-off-by: yangshihao6 <[email protected]>
Model module is useless now. Let't remove it totally. - vLLM version: v0.11.2 Signed-off-by: yangshihao6 <[email protected]>
Model module is useless now. Let't remove it totally. - vLLM version: v0.11.2 Signed-off-by: yangshihao6 <[email protected]>
# Conflicts: # docs/source/tutorials/Qwen2.5.md
Model module is useless now. Let't remove it totally. - vLLM version: v0.11.2 Signed-off-by: yangshihao6 <[email protected]>
What this PR does / why we need it?
Does this PR introduce any user-facing change?
How was this patch tested?