diff --git a/convert-hf-to-gguf.py b/convert-hf-to-gguf.py index 025405a2c6ce1..c7a66afaecdfa 100755 --- a/convert-hf-to-gguf.py +++ b/convert-hf-to-gguf.py @@ -2191,12 +2191,18 @@ def phantom(tok): def modify_tensors(self, data_torch: Tensor, name: str, bid: int | None) -> Iterable[tuple[str, Tensor]]: del bid # unused - + + name = name.removeprefix("bert.") + # we are only using BERT for embeddings so we don't need the pooling layer - if name in ("embeddings.position_ids", "pooler.dense.weight", "pooler.dense.bias"): + if name in ("embeddings.position_ids", "pooler.dense.weight", "pooler.dense.bias") or "cls." in name: return [] # we don't need these - return [(self.map_tensor_name(name), data_torch)] + name = name.replace("gamma", "weight") + name = name.replace("beta", "bias") + + try_suffixes = (".weight", ".bias", ".beta", ".gamma") + return [(self.map_tensor_name(name, try_suffixes), data_torch)] @Model.register("NomicBertModel")