Skip to content

Commit 209f72b

Browse files
committed
allow to keep FP32/FP16/BF16 default hf to gguf
1 parent f918e3f commit 209f72b

File tree

1 file changed

+7
-3
lines changed

1 file changed

+7
-3
lines changed

convert_hf_to_gguf.py

+7-3
Original file line numberDiff line numberDiff line change
@@ -100,13 +100,17 @@ def __init__(self, dir_model: Path, ftype: gguf.LlamaFileType, fname_out: Path,
100100
if self.ftype == gguf.LlamaFileType.GUESSED:
101101
# NOTE: can't use field "torch_dtype" in config.json, because some finetunes lie.
102102
_, first_tensor = next(self.get_tensors())
103-
if first_tensor.dtype == torch.float16:
103+
if first_tensor.dtype == torch.float32:
104+
logger.info(f"choosing --outtype f32 from first tensor type ({first_tensor.dtype})")
105+
self.ftype = gguf.LlamaFileType.ALL_F32
106+
elif first_tensor.dtype == torch.float16:
104107
logger.info(f"choosing --outtype f16 from first tensor type ({first_tensor.dtype})")
105108
self.ftype = gguf.LlamaFileType.MOSTLY_F16
106-
else:
109+
elif first_tensor.dtype == torch.bfloat16:
107110
logger.info(f"choosing --outtype bf16 from first tensor type ({first_tensor.dtype})")
108111
self.ftype = gguf.LlamaFileType.MOSTLY_BF16
109-
112+
else:
113+
raise TypeError(f"can't guessed tensor type: first tensor type ({first_tensor.dtype})")
110114
# Configure GGUF Writer
111115
self.gguf_writer = gguf.GGUFWriter(path=None, arch=gguf.MODEL_ARCH_NAMES[self.model_arch], endianess=self.endianess, use_temp_file=self.use_temp_file,
112116
split_max_tensors=split_max_tensors, split_max_size=split_max_size, dry_run=dry_run, small_first_shard=small_first_shard)

0 commit comments

Comments
 (0)