Skip to content

Commit 1dd058c

Browse files
committed
chore: update docs
Signed-off-by: Aaron Pham <contact@aarnphm.xyz>
1 parent c0fa317 commit 1dd058c

File tree

1 file changed

+22
-24
lines changed

1 file changed

+22
-24
lines changed

docs/source/features/reasoning_outputs.md

Lines changed: 22 additions & 24 deletions
Original file line numberDiff line numberDiff line change
@@ -10,11 +10,11 @@ Reasoning models return an additional `reasoning_content` field in their outputs
1010

1111
vLLM currently supports the following reasoning models:
1212

13-
| Model Series | Parser Name | Structured Output Support | Tool Calling |
14-
|--------------|-------------|------------------|-------------|
15-
| [DeepSeek R1 series](https://huggingface.co/collections/deepseek-ai/deepseek-r1-678e1e131c0169c0bc89728d) | `deepseek_r1` | `guided_json`, `guided_regex` ||
16-
| [QwQ-32B](https://huggingface.co/Qwen/QwQ-32B) | `deepseek_r1` | `guided_json`, `guided_regex` ||
17-
| [IBM Granite 3.2 language models](https://huggingface.co/collections/ibm-granite/granite-32-language-models-67b3bc8c13508f6d064cff9a) | `granite` |||
13+
| Model Series | Parser Name | Structured Output Support | Tool Calling |
14+
| ------------------------------------------------------------------------------------------------------------------------------------- | ------------- | ----------------------------- | ------------ |
15+
| [DeepSeek R1 series](https://huggingface.co/collections/deepseek-ai/deepseek-r1-678e1e131c0169c0bc89728d) | `deepseek_r1` | `guided_json`, `guided_regex` | |
16+
| [QwQ-32B](https://huggingface.co/Qwen/QwQ-32B) | `deepseek_r1` | `guided_json`, `guided_regex` | |
17+
| [IBM Granite 3.2 language models](https://huggingface.co/collections/ibm-granite/granite-32-language-models-67b3bc8c13508f6d064cff9a) | `granite` | | |
1818

1919
- IBM Granite 3.2 reasoning is disabled by default; to enable it, you must also pass `thinking=True` in your `chat_template_kwargs`.
2020

@@ -64,22 +64,22 @@ Streaming chat completions are also supported for reasoning models. The `reasoni
6464

6565
```json
6666
{
67-
"id": "chatcmpl-123",
68-
"object": "chat.completion.chunk",
69-
"created": 1694268190,
70-
"model": "deepseek-ai/DeepSeek-R1-Distill-Qwen-1.5B",
71-
"system_fingerprint": "fp_44709d6fcb",
72-
"choices": [
73-
{
74-
"index": 0,
75-
"delta": {
76-
"role": "assistant",
77-
"reasoning_content": "is",
78-
},
79-
"logprobs": null,
80-
"finish_reason": null
81-
}
82-
]
67+
"id": "chatcmpl-123",
68+
"object": "chat.completion.chunk",
69+
"created": 1694268190,
70+
"model": "deepseek-ai/DeepSeek-R1-Distill-Qwen-1.5B",
71+
"system_fingerprint": "fp_44709d6fcb",
72+
"choices": [
73+
{
74+
"index": 0,
75+
"delta": {
76+
"role": "assistant",
77+
"reasoning_content": "is"
78+
},
79+
"logprobs": null,
80+
"finish_reason": null
81+
}
82+
]
8383
}
8484
```
8585

@@ -139,12 +139,10 @@ Remember to check whether the `reasoning_content` exists in the response before
139139
The reasoning content is also available in the structured output. The structured output engine like `xgrammar` will use the reasoning content to generate structured output. It is only supported in v0 engine now.
140140

141141
```bash
142-
VLLM_USE_V1=0 vllm serve deepseek-ai/DeepSeek-R1-Distill-Qwen-1.5B \
142+
vllm serve deepseek-ai/DeepSeek-R1-Distill-Qwen-1.5B \
143143
--enable-reasoning --reasoning-parser deepseek_r1
144144
```
145145

146-
Please note that the `VLLM_USE_V1` environment variable must be set to `0` to use the v0 engine.
147-
148146
```python
149147
from openai import OpenAI
150148
from pydantic import BaseModel

0 commit comments

Comments
 (0)