We read every piece of feedback, and take your input very seriously.
To see all available qualifiers, see our documentation.
There was an error while loading. Please reload this page.
1 parent 2b79376 commit 7dfddcdCopy full SHA for 7dfddcd
vllm/model_executor/models/llama.py
@@ -28,6 +28,7 @@
28
from transformers import LlamaConfig
29
30
from vllm.attention import Attention, AttentionMetadata
31
+from vllm.compilation.decorators import support_compile_llama_style
32
from vllm.config import CacheConfig, LoRAConfig
33
from vllm.distributed import (get_pp_group, get_tensor_model_parallel_rank,
34
get_tensor_model_parallel_world_size)
@@ -344,6 +345,7 @@ def forward(
344
345
return hidden_states
346
347
348
+@support_compile_llama_style
349
class LlamaForCausalLM(nn.Module, SupportsLoRA):
350
packed_modules_mapping = {
351
"qkv_proj": [
0 commit comments