Skip to content

Commit ac5520c

Browse files
committed
add asserts back
1 parent a33ea39 commit ac5520c

File tree

2 files changed

+14
-0
lines changed

2 files changed

+14
-0
lines changed

vllm/entrypoints/openai/serving_chat.py

Lines changed: 7 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -9,6 +9,8 @@
99
from fastapi import Request
1010

1111
from vllm.config import ModelConfig
12+
from vllm.engine.async_llm_engine import AsyncLLMEngine
13+
from vllm.engine.multiprocessing.client import MQLLMEngineClient
1214
from vllm.engine.protocol import EngineClient
1315
from vllm.entrypoints.chat_utils import (ConversationMessage,
1416
apply_hf_chat_template,
@@ -235,6 +237,11 @@ async def create_chat_completion(
235237
log_tracing_disabled_warning()
236238

237239
if isinstance(sampling_params, BeamSearchParams):
240+
assert isinstance(self.engine_client,
241+
(AsyncLLMEngine,
242+
MQLLMEngineClient)), \
243+
"Beam search is only supported with" \
244+
"AsyncLLMEngine and MQLLMEngineClient."
238245
result_generator = self.engine_client.beam_search(
239246
engine_inputs['prompt_token_ids'],
240247
request_id,

vllm/entrypoints/openai/serving_completion.py

Lines changed: 7 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -8,6 +8,8 @@
88
from fastapi import Request
99

1010
from vllm.config import ModelConfig
11+
from vllm.engine.async_llm_engine import AsyncLLMEngine
12+
from vllm.engine.multiprocessing.client import MQLLMEngineClient
1113
from vllm.engine.protocol import EngineClient
1214
from vllm.entrypoints.logger import RequestLogger
1315
# yapf conflicts with isort for this block
@@ -149,6 +151,11 @@ async def create_completion(
149151
log_tracing_disabled_warning()
150152

151153
if isinstance(sampling_params, BeamSearchParams):
154+
assert isinstance(self.engine_client,
155+
(AsyncLLMEngine,
156+
MQLLMEngineClient)), \
157+
"Beam search is only supported with" \
158+
"AsyncLLMEngine and MQLLMEngineClient."
152159
generator = self.engine_client.beam_search(
153160
prompt_inputs["prompt_token_ids"],
154161
request_id,

0 commit comments

Comments
 (0)