@@ -296,7 +296,9 @@ def byref(obj: CtypesCData, offset: Optional[int] = None) -> CtypesRef[CtypesCDa
296
296
# LLAMA_VOCAB_PRE_TYPE_GPT2 = 7,
297
297
# LLAMA_VOCAB_PRE_TYPE_REFACT = 8,
298
298
# LLAMA_VOCAB_PRE_TYPE_COMMAND_R = 9,
299
- # LLAMA_VOCAB_PRE_TYPE_OLMO = 10,
299
+ # LLAMA_VOCAB_PRE_TYPE_QWEN2 = 10,
300
+ # LLAMA_VOCAB_PRE_TYPE_OLMO = 11,
301
+ # LLAMA_VOCAB_PRE_TYPE_DBRX = 12,
300
302
# };
301
303
LLAMA_VOCAB_PRE_TYPE_DEFAULT = 0
302
304
LLAMA_VOCAB_PRE_TYPE_LLAMA3 = 1
@@ -308,7 +310,9 @@ def byref(obj: CtypesCData, offset: Optional[int] = None) -> CtypesRef[CtypesCDa
308
310
LLAMA_VOCAB_PRE_TYPE_GPT2 = 7
309
311
LLAMA_VOCAB_PRE_TYPE_REFACT = 8
310
312
LLAMA_VOCAB_PRE_TYPE_COMMAND_R = 9
311
- LLAMA_VOCAB_PRE_TYPE_OLMO = 10
313
+ LLAMA_VOCAB_PRE_TYPE_QWEN2 = 10
314
+ LLAMA_VOCAB_PRE_TYPE_OLMO = 11
315
+ LLAMA_VOCAB_PRE_TYPE_DBRX = 12
312
316
313
317
314
318
# // note: these values should be synchronized with ggml_rope
@@ -377,6 +381,7 @@ def byref(obj: CtypesCData, offset: Optional[int] = None) -> CtypesRef[CtypesCDa
377
381
# LLAMA_FTYPE_MOSTLY_IQ2_M = 29, // except 1d tensors
378
382
# LLAMA_FTYPE_MOSTLY_IQ4_XS = 30, // except 1d tensors
379
383
# LLAMA_FTYPE_MOSTLY_IQ1_M = 31, // except 1d tensors
384
+ # LLAMA_FTYPE_MOSTLY_BF16 = 32, // except 1d tensors
380
385
381
386
# LLAMA_FTYPE_GUESSED = 1024, // not specified in the model file
382
387
# };
@@ -409,6 +414,8 @@ def byref(obj: CtypesCData, offset: Optional[int] = None) -> CtypesRef[CtypesCDa
409
414
LLAMA_FTYPE_MOSTLY_IQ2_S = 28
410
415
LLAMA_FTYPE_MOSTLY_IQ2_M = 29
411
416
LLAMA_FTYPE_MOSTLY_IQ4_XS = 30
417
+ LLAMA_FTYPE_MOSTLY_IQ1_M = 31
418
+ LLAMA_FTYPE_MOSTLY_BF16 = 32
412
419
LLAMA_FTYPE_GUESSED = 1024
413
420
414
421
# enum llama_rope_scaling_type {
0 commit comments