Skip to content

Commit 7dfddcd

Browse files
committed
support llama
1 parent 2b79376 commit 7dfddcd

File tree

1 file changed

+2
-0
lines changed

1 file changed

+2
-0
lines changed

vllm/model_executor/models/llama.py

Lines changed: 2 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -28,6 +28,7 @@
2828
from transformers import LlamaConfig
2929

3030
from vllm.attention import Attention, AttentionMetadata
31+
from vllm.compilation.decorators import support_compile_llama_style
3132
from vllm.config import CacheConfig, LoRAConfig
3233
from vllm.distributed import (get_pp_group, get_tensor_model_parallel_rank,
3334
get_tensor_model_parallel_world_size)
@@ -344,6 +345,7 @@ def forward(
344345
return hidden_states
345346

346347

348+
@support_compile_llama_style
347349
class LlamaForCausalLM(nn.Module, SupportsLoRA):
348350
packed_modules_mapping = {
349351
"qkv_proj": [

0 commit comments

Comments
 (0)