Skip to content

Commit af59780

Browse files
committed
Fix unit tests reference on served_model_names
1 parent c454206 commit af59780

File tree

3 files changed

+11
-4
lines changed

3 files changed

+11
-4
lines changed

tests/entrypoints/openai/test_serving_chat.py

Lines changed: 5 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -7,11 +7,14 @@
77
from vllm.engine.async_llm_engine import AsyncLLMEngine
88
from vllm.entrypoints.openai.protocol import ChatCompletionRequest
99
from vllm.entrypoints.openai.serving_chat import OpenAIServingChat
10+
from vllm.entrypoints.openai.serving_engine import BaseModelPath
1011
from vllm.transformers_utils.tokenizer import get_tokenizer
1112

1213
MODEL_NAME = "openai-community/gpt2"
1314
CHAT_TEMPLATE = "Dummy chat template for testing {}"
1415

16+
BASE_MODEL_PATHS = [BaseModelPath(name=MODEL_NAME, model_path=MODEL_NAME)]
17+
1518

1619
@dataclass
1720
class MockModelConfig:
@@ -37,7 +40,7 @@ async def _async_serving_chat_init():
3740

3841
serving_completion = OpenAIServingChat(engine,
3942
model_config,
40-
served_model_names=[MODEL_NAME],
43+
BASE_MODEL_PATHS,
4144
response_role="assistant",
4245
chat_template=CHAT_TEMPLATE,
4346
lora_modules=None,
@@ -57,7 +60,7 @@ def test_serving_chat_should_set_correct_max_tokens():
5760

5861
serving_chat = OpenAIServingChat(mock_engine,
5962
MockModelConfig(),
60-
served_model_names=[MODEL_NAME],
63+
BASE_MODEL_PATHS,
6164
response_role="assistant",
6265
chat_template=CHAT_TEMPLATE,
6366
lora_modules=None,

vllm/entrypoints/openai/run_batch.py

Lines changed: 6 additions & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -17,6 +17,7 @@
1717
# yapf: enable
1818
from vllm.entrypoints.openai.serving_chat import OpenAIServingChat
1919
from vllm.entrypoints.openai.serving_embedding import OpenAIServingEmbedding
20+
from vllm.entrypoints.openai.serving_engine import BaseModelPath
2021
from vllm.usage.usage_lib import UsageContext
2122
from vllm.utils import FlexibleArgumentParser, random_uuid
2223
from vllm.version import __version__ as VLLM_VERSION
@@ -140,6 +141,10 @@ async def main(args):
140141

141142
# When using single vLLM without engine_use_ray
142143
model_config = await engine.get_model_config()
144+
base_model_paths = [
145+
BaseModelPath(name=name, model_path=args.model)
146+
for name in served_model_names
147+
]
143148

144149
if args.disable_log_requests:
145150
request_logger = None
@@ -150,7 +155,7 @@ async def main(args):
150155
openai_serving_chat = OpenAIServingChat(
151156
engine,
152157
model_config,
153-
served_model_names,
158+
base_model_paths,
154159
args.response_role,
155160
lora_modules=None,
156161
prompt_adapters=None,

vllm/entrypoints/openai/serving_engine.py

Lines changed: 0 additions & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -495,4 +495,3 @@ async def unload_lora_adapter(
495495

496496
def _is_model_supported(self, model_name):
497497
return any(model.name == model_name for model in self.base_model_paths)
498-

0 commit comments

Comments
 (0)