Skip to content

Commit 2ed6ffc

Browse files
committed
link PR number
Signed-off-by: KuntaiDu <kuntai@uchicago.edu>
1 parent e5530f0 commit 2ed6ffc

File tree

1 file changed

+1
-1
lines changed

1 file changed

+1
-1
lines changed

vllm/engine/arg_utils.py

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -1463,7 +1463,7 @@ def _set_default_args_v1(self, usage_context: UsageContext) -> None:
14631463
device_memory = 0
14641464

14651465
# NOTE(Kuntai): Setting large `max_num_batched_tokens` for A100 reduces
1466-
# throughput, see PR # ??? for more details.
1466+
# throughput, see PR #17885 for more details.
14671467
# So here we do an extra device name check to prevent such regression.
14681468
if device_memory >= 70 * GiB_bytes and "a100" not in device_name:
14691469
# For GPUs like H100 and MI300x, use larger default values.

0 commit comments

Comments
 (0)