Skip to content

[Bug]: 使用vllm-ascend的v0.11.0rc1镜像跑不通v0.11.0rc1分支的一个脚本,但是0.9.1镜像能跑通,脚本路径:https://github.com/vllm-project/vllm/blob/v0.11.0rc1/examples/offline_inference/prompt_embed_inference.py #4171

@hwj-cl

Description

@hwj-cl

Your current environment

A2服务器上

🐛 Describe the bug

0.11.0rc1镜像:跑prompt_embed_inference.py脚本时,在最开始初始化的时候卡住,查看日志没有报错,NPU显存占用117左右;

0.9.1镜像:跑prompt_embed_inference.py通过,有类似以下输出
[Single Inference Output]

The capital of France is Paris. It is located in the northwest of the country

[Batch Inference Outputs]

Q1:Please tell me about the capital of France,
Al:The capital of France is Paris,It is located in the 1le-de-F

A2: The day with the longest duration in a year is typically the summer solstice,
Q2:When is the day longest during the year?

Q3:Where is bigger, the moon or the sum?
A3: The sun is significantly larger than the moon Here are some key facts to compare

Metadata

Metadata

Assignees

No one assigned

    Labels

    bugSomething isn't working

    Type

    No type

    Projects

    No projects

    Milestone

    No milestone

    Relationships

    None yet

    Development

    No branches or pull requests

    Issue actions