Skip to content

Commit 5166aaf

Browse files
authored
convert : force patch_merger tensors to f16/f32 (#18124)
1 parent 6ce3d85 commit 5166aaf

File tree

1 file changed

+1
-1
lines changed

1 file changed

+1
-1
lines changed

convert_hf_to_gguf.py

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -1838,7 +1838,7 @@ def _find_param(self, obj: dict[str, Any], keys: Iterable[str], optional: bool =
18381838

18391839
def tensor_force_quant(self, name, new_name, bid, n_dims):
18401840
del bid, name, n_dims # unused
1841-
if ".patch_embd.weight" in new_name:
1841+
if ".patch_embd.weight" in new_name or ".patch_merger.weight" in new_name:
18421842
return gguf.GGMLQuantizationType.F16 if self.ftype == gguf.LlamaFileType.MOSTLY_F16 else gguf.GGMLQuantizationType.F32
18431843
return False
18441844

0 commit comments

Comments
 (0)