Skip to content

Commit 742f679

Browse files
authored
Remove prompt string from engine core data structures (#663)
### What this PR does / why we need it? vLLM Ascend side followup on: [Core] Remove prompt string from engine core data structures vllm-project/vllm@df6f3ce ### Does this PR introduce _any_ user-facing change? No ### How was this patch tested? CI passed Signed-off-by: Yikun Jiang <[email protected]>
1 parent c99c4c8 commit 742f679

File tree

1 file changed

+28
-14
lines changed

1 file changed

+28
-14
lines changed

vllm_ascend/worker/model_runner_v1.py

Lines changed: 28 additions & 14 deletions
Original file line numberDiff line numberDiff line change
@@ -346,20 +346,34 @@ def _update_states(self, scheduler_output: "SchedulerOutput") -> None:
346346
generator.manual_seed(sampling_params.seed)
347347
else:
348348
generator = None
349-
350-
self.requests[req_id] = CachedRequestState(
351-
req_id=req_id,
352-
prompt_token_ids=new_req_data.prompt_token_ids,
353-
prompt=new_req_data.prompt,
354-
mm_inputs=new_req_data.mm_inputs,
355-
mm_positions=new_req_data.mm_positions,
356-
sampling_params=sampling_params,
357-
generator=generator,
358-
block_ids=new_req_data.block_ids,
359-
num_computed_tokens=new_req_data.num_computed_tokens,
360-
output_token_ids=[],
361-
lora_request=new_req_data.lora_request,
362-
)
349+
if vllm_version_is("0.8.4"):
350+
self.requests[req_id] = CachedRequestState(
351+
req_id=req_id,
352+
prompt_token_ids=new_req_data.prompt_token_ids,
353+
prompt=new_req_data.prompt,
354+
mm_inputs=new_req_data.mm_inputs,
355+
mm_positions=new_req_data.mm_positions,
356+
sampling_params=sampling_params,
357+
generator=generator,
358+
block_ids=new_req_data.block_ids,
359+
num_computed_tokens=new_req_data.num_computed_tokens,
360+
output_token_ids=[],
361+
lora_request=new_req_data.lora_request,
362+
)
363+
else:
364+
# the prompt removed by: https://github.com/vllm-project/vllm/pull/17214
365+
self.requests[req_id] = CachedRequestState(
366+
req_id=req_id,
367+
prompt_token_ids=new_req_data.prompt_token_ids,
368+
mm_inputs=new_req_data.mm_inputs,
369+
mm_positions=new_req_data.mm_positions,
370+
sampling_params=sampling_params,
371+
generator=generator,
372+
block_ids=new_req_data.block_ids,
373+
num_computed_tokens=new_req_data.num_computed_tokens,
374+
output_token_ids=[],
375+
lora_request=new_req_data.lora_request,
376+
)
363377

364378
req_ids_to_add.append(req_id)
365379

0 commit comments

Comments
 (0)