File tree 2 files changed +8
-1
lines changed
2 files changed +8
-1
lines changed Original file line number Diff line number Diff line change @@ -322,6 +322,9 @@ tokenizer_type: AutoTokenizer
322
322
trust_remote_code:
323
323
# use_fast option for tokenizer loading from_pretrained, default to True
324
324
tokenizer_use_fast:
325
+ # resize the model embeddings when new tokens are added to multiples of 32
326
+ # this is reported to improve training speed on some models
327
+ resize_token_embeddings_to_32x:
325
328
326
329
# whether you are training a 4-bit GPTQ quantized model
327
330
gptq: true
Original file line number Diff line number Diff line change @@ -301,7 +301,11 @@ def load_model(
301
301
** model_kwargs ,
302
302
)
303
303
304
- embeddings_len = math .ceil (len (tokenizer ) / 32 ) * 32
304
+ embeddings_len = (
305
+ math .ceil (len (tokenizer ) / 32 ) * 32
306
+ if cfg .resize_token_embeddings_to_32x
307
+ else len (tokenizer )
308
+ )
305
309
model .resize_token_embeddings (embeddings_len )
306
310
307
311
if (
You can’t perform that action at this time.
0 commit comments