diff options
author | Jared Van Bortel <jared@nomic.ai> | 2024-02-23 13:39:14 -0500 |
---|---|---|
committer | GitHub <noreply@github.com> | 2024-02-23 20:39:14 +0200 |
commit | 54fbcd2ce6c48c9e22eca6fbf9e53fb68c3e72ea (patch) | |
tree | b944748a597fd6ab7813f3c823863e87bfa858a7 /convert-hf-to-gguf.py | |
parent | 15499eb94227401bdc8875da6eb85c15d37068f7 (diff) |
convert : fix missing ftype for gemma (#5690)
Diffstat (limited to 'convert-hf-to-gguf.py')
-rwxr-xr-x | convert-hf-to-gguf.py | 1 |
1 files changed, 1 insertions, 0 deletions
diff --git a/convert-hf-to-gguf.py b/convert-hf-to-gguf.py index 9bdfce07..32d54b45 100755 --- a/convert-hf-to-gguf.py +++ b/convert-hf-to-gguf.py @@ -1803,6 +1803,7 @@ class GemmaModel(Model): self.gguf_writer.add_layer_norm_rms_eps(self.hparams["rms_norm_eps"]) self.gguf_writer.add_key_length(hparams["head_dim"]) self.gguf_writer.add_value_length(hparams["head_dim"]) + self.gguf_writer.add_file_type(self.ftype) def write_tensors(self): block_count = self.hparams.get("n_layers", self.hparams.get("num_hidden_layers", self.hparams.get("n_layer"))) |