Skip to content

Commit 4746c84

Browse files
committed
Update llama-quant.cpp
1 parent ed670d9 commit 4746c84

File tree

1 file changed

+2
-2
lines changed

1 file changed

+2
-2
lines changed

src/llama-quant.cpp

+2-2
Original file line numberDiff line numberDiff line change
@@ -360,13 +360,13 @@ static ggml_type llama_tensor_get_type(quantize_state_impl & qs, ggml_type new_t
360360
// if (use_more_bits(i_layer, n_layer)) new_type = GGML_TYPE_Q6_K;
361361
// }
362362
// }
363-
else if (ftype == LLAMA_FTYPE_MOSTLY_Q4_K_S && use_more_bits(i_layer, n_layer)) new_type = GGML_TYPE_Q5_K;
363+
// else if (ftype == LLAMA_FTYPE_MOSTLY_Q4_K_S && use_more_bits(i_layer, n_layer)) new_type = GGML_TYPE_Q5_K;
364364
else if (ftype == LLAMA_FTYPE_MOSTLY_Q4_K_M) new_type = GGML_TYPE_Q5_K;
365365
else if (i_layer < n_layer/8 && (ftype == LLAMA_FTYPE_MOSTLY_IQ4_NL || ftype == LLAMA_FTYPE_MOSTLY_IQ4_XS)
366366
&& !qs.has_imatrix && use_more_bits(i_layer, n_layer)) {
367367
new_type = GGML_TYPE_Q5_K;
368368
}
369-
else if (ftype == LLAMA_FTYPE_MOSTLY_Q5_K_S && use_more_bits(i_layer, n_layer)) new_type = GGML_TYPE_Q6_K;
369+
// else if (ftype == LLAMA_FTYPE_MOSTLY_Q5_K_S && use_more_bits(i_layer, n_layer)) new_type = GGML_TYPE_Q6_K;
370370
else if (ftype == LLAMA_FTYPE_MOSTLY_Q5_K_M) new_type = GGML_TYPE_Q6_K;
371371
else if (ftype == LLAMA_FTYPE_MOSTLY_Q4_K_S && arch != LLM_ARCH_FALCON && i_layer < n_layer/8) {
372372
new_type = GGML_TYPE_Q5_K;

0 commit comments

Comments
 (0)