File tree Expand file tree Collapse file tree 1 file changed +4
-3
lines changed
vllm/model_executor/models Expand file tree Collapse file tree 1 file changed +4
-3
lines changed Original file line number Diff line number Diff line change @@ -257,8 +257,8 @@ def __init__(
257
257
(lora_config .max_loras or 1 )) if lora_config else 0
258
258
self .vocab_size = config .vocab_size + lora_vocab
259
259
self .org_vocab_size = config .vocab_size
260
- if get_pp_group ().is_first_rank or (config .tie_word_embeddings and
261
- get_pp_group ().is_last_rank ):
260
+ if get_pp_group ().is_first_rank or (config .tie_word_embeddings
261
+ and get_pp_group ().is_last_rank ):
262
262
self .embed_tokens = VocabParallelEmbedding (
263
263
self .vocab_size ,
264
264
config .hidden_size ,
@@ -386,7 +386,8 @@ def __init__(
386
386
387
387
logit_scale = getattr (config , "logit_scale" , 1.0 )
388
388
self .logits_processor = LogitsProcessor (self .unpadded_vocab_size ,
389
- config .vocab_size , logit_scale )
389
+ config .vocab_size ,
390
+ logit_scale )
390
391
self .sampler = Sampler ()
391
392
else :
392
393
self .lm_head = PPMissingLayer ()
You can’t perform that action at this time.
0 commit comments