We read every piece of feedback, and take your input very seriously.
To see all available qualifiers, see our documentation.
There was an error while loading. Please reload this page.
1 parent 069369f commit 225ec48Copy full SHA for 225ec48
convert-hf-to-gguf.py
@@ -296,7 +296,7 @@ def write_tensors(self):
296
if self.ftype != gguf.LlamaFileType.ALL_F32 and extra_f16 and not extra_f32:
297
if self.ftype == gguf.LlamaFileType.MOSTLY_BF16:
298
data = gguf.truncate_bf16(data) if old_dtype == torch.bfloat16 else gguf.quantize_bf16(data)
299
- assert data.dtype in (np.int16, np.uint16)
+ assert data.dtype == np.uint16
300
data_qtype = gguf.GGMLQuantizationType.BF16
301
302
elif self.ftype == gguf.LlamaFileType.MOSTLY_Q8_0 and gguf.can_quantize_to_q8_0(data):
0 commit comments