You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
I ran the first test function in vllm/tests/v1/engine/test_engine_core.py, and modified the EngineArgs by adding the parameters pipeline_parallel_size=2 and distributed_executor_backend='ray'. However, the error message shows that FutureWrapper has no attribute sampled_token_ids.
At the same time, I modified the update_from_output method of the Scheduler class in vllm/v1/core/sched/scheduler.py, trying to extract result from model_runner_output, but this did not solve the problem. When running the test code, I encountered another error: 'ModelRunnerOutput' object has no attribute 'finished_req_ids'. It seems like somewhere inside Ray, a SchedulerOutput was expected, but a ModelRunnerOutput was actually passed.
Below, I’ve included the two error messages and the function I used from the test folder.
Traceback (most recent call last):
File "/data/user/test/vllm/tests/utils.py", line 727, in wrapper
f(*args, **kwargs)
File "/data/user/test/vllm/tests/v1/engine/test_engine_core.py", line 67, in test_engine_core
_ = engine_core.step()
File "/data/conda_envs/sllm-store/lib/python3.10/site-packages/vllm/v1/engine/core.py", line 208, in step
engine_core_outputs = self.scheduler.update_from_output(
File "/data/conda_envs/sllm-store/lib/python3.10/site-packages/vllm/v1/core/sched/scheduler.py", line 630, in update_from_output
sampled_token_ids = model_runner_output.sampled_token_ids
AttributeError: 'FutureWrapper' object has no attribute 'sampled_token_ids'
Traceback (most recent call last):
File "/data/user/test/vllm/tests/utils.py", line 727, in wrapper
f(args, **kwargs)
File "/data/user/test/vllm/tests/v1/engine/test_engine_core.py", line 91, in test_engine_core
while len(engine_core.step().outputs) > 0:
File "/data/conda_envs/sllm-store/lib/python3.10/site-packages/vllm/v1/engine/core.py", line 208, in step
engine_core_outputs = self.scheduler.update_from_output(
File "/data/conda_envs/sllm-store/lib/python3.10/site-packages/vllm/v1/core/sched/scheduler.py", line 630, in update_from_output
model_runner_output = model_runner_output.result()
File "/data/conda_envs/sllm-store/lib/python3.10/site-packages/vllm/v1/executor/ray_distributed_executor.py", line 24, in result
return self.ref.get()
File "/data/conda_envs/sllm-store/lib/python3.10/site-packages/ray/experimental/compiled_dag_ref.py", line 150, in get
return _process_return_vals(return_vals, True)
File "/data/conda_envs/sllm-store/lib/python3.10/site-packages/ray/experimental/compiled_dag_ref.py", line 27, in _process_return_vals
raise val.as_instanceof_cause()
ray.exceptions.RayTaskError(AttributeError): ray::RayWorkerWrapper.ray_call() (pid=2675487, ip=)
File "/data/conda_envs/sllm-store/lib/python3.10/site-packages/vllm/executor/ray_utils.py", line 139, in execute_model_ray
output = self.worker.model_runner.execute_model(
File "/data/conda_envs/sllm-store/lib/python3.10/site-packages/torch/utils/_contextlib.py", line 116, in decorate_context
return func(*args, **kwargs)
File "/data/conda_envs/sllm-store/lib/python3.10/site-packages/vllm/v1/worker/gpu_model_runner.py", line 1012, in execute_model
self._update_states(scheduler_output)
File "/data/conda_envs/sllm-store/lib/python3.10/site-packages/vllm/v1/worker/gpu_model_runner.py", line 296, in _update_states
for req_id in scheduler_output.finished_req_ids:
AttributeError: 'ModelRunnerOutput' object has no attribute 'finished_req_ids'
Before submitting a new issue...
Make sure you already searched for relevant issues, and asked the chatbot living at the bottom right corner of the documentation page, which can answer lots of frequently asked questions.
The text was updated successfully, but these errors were encountered:
What version of vLLM are you using? Have you tried the latest version?
Thanks for the reminder. I just tried the latest version 0.9.1, but the issue still persists. From what I can see, it seems like 0.9.1 doesn’t fix the pp bug yet?
Uh oh!
There was an error while loading. Please reload this page.
Your current environment
The output of
python collect_env.py
🐛 Describe the bug
I ran the first test function in vllm/tests/v1/engine/test_engine_core.py, and modified the EngineArgs by adding the parameters pipeline_parallel_size=2 and distributed_executor_backend='ray'. However, the error message shows that FutureWrapper has no attribute sampled_token_ids.
At the same time, I modified the update_from_output method of the Scheduler class in vllm/v1/core/sched/scheduler.py, trying to extract result from model_runner_output, but this did not solve the problem. When running the test code, I encountered another error: 'ModelRunnerOutput' object has no attribute 'finished_req_ids'. It seems like somewhere inside Ray, a SchedulerOutput was expected, but a ModelRunnerOutput was actually passed.
Below, I’ve included the two error messages and the function I used from the test folder.
def test_engine_core(monkeypatch: pytest.MonkeyPatch):
Traceback (most recent call last):
File "/data/user/test/vllm/tests/utils.py", line 727, in wrapper
f(*args, **kwargs)
File "/data/user/test/vllm/tests/v1/engine/test_engine_core.py", line 67, in test_engine_core
_ = engine_core.step()
File "/data/conda_envs/sllm-store/lib/python3.10/site-packages/vllm/v1/engine/core.py", line 208, in step
engine_core_outputs = self.scheduler.update_from_output(
File "/data/conda_envs/sllm-store/lib/python3.10/site-packages/vllm/v1/core/sched/scheduler.py", line 630, in update_from_output
sampled_token_ids = model_runner_output.sampled_token_ids
AttributeError: 'FutureWrapper' object has no attribute 'sampled_token_ids'
Traceback (most recent call last):
File "/data/user/test/vllm/tests/utils.py", line 727, in wrapper
f(args, **kwargs)
File "/data/user/test/vllm/tests/v1/engine/test_engine_core.py", line 91, in test_engine_core
while len(engine_core.step().outputs) > 0:
File "/data/conda_envs/sllm-store/lib/python3.10/site-packages/vllm/v1/engine/core.py", line 208, in step
engine_core_outputs = self.scheduler.update_from_output(
File "/data/conda_envs/sllm-store/lib/python3.10/site-packages/vllm/v1/core/sched/scheduler.py", line 630, in update_from_output
model_runner_output = model_runner_output.result()
File "/data/conda_envs/sllm-store/lib/python3.10/site-packages/vllm/v1/executor/ray_distributed_executor.py", line 24, in result
return self.ref.get()
File "/data/conda_envs/sllm-store/lib/python3.10/site-packages/ray/experimental/compiled_dag_ref.py", line 150, in get
return _process_return_vals(return_vals, True)
File "/data/conda_envs/sllm-store/lib/python3.10/site-packages/ray/experimental/compiled_dag_ref.py", line 27, in _process_return_vals
raise val.as_instanceof_cause()
ray.exceptions.RayTaskError(AttributeError): ray::RayWorkerWrapper.ray_call() (pid=2675487, ip=)
File "/data/conda_envs/sllm-store/lib/python3.10/site-packages/vllm/executor/ray_utils.py", line 139, in execute_model_ray
output = self.worker.model_runner.execute_model(
File "/data/conda_envs/sllm-store/lib/python3.10/site-packages/torch/utils/_contextlib.py", line 116, in decorate_context
return func(*args, **kwargs)
File "/data/conda_envs/sllm-store/lib/python3.10/site-packages/vllm/v1/worker/gpu_model_runner.py", line 1012, in execute_model
self._update_states(scheduler_output)
File "/data/conda_envs/sllm-store/lib/python3.10/site-packages/vllm/v1/worker/gpu_model_runner.py", line 296, in _update_states
for req_id in scheduler_output.finished_req_ids:
AttributeError: 'ModelRunnerOutput' object has no attribute 'finished_req_ids'
Before submitting a new issue...
The text was updated successfully, but these errors were encountered: