diff options
author | David Sommers <12738+databyte@users.noreply.github.com> | 2024-01-18 12:20:59 -0500 |
---|---|---|
committer | GitHub <noreply@github.com> | 2024-01-18 19:20:59 +0200 |
commit | b46757735d30f5c6ed4f20ebeccc684e02d4f3bf (patch) | |
tree | 7327dd01fecf2f45007ae84e9d09345f317083bb | |
parent | 3e945cc1e9c06d2001031360e4e303e9548fb02c (diff) |
convert.py : fix llama/llama2 conversion due to vocab_size=-1 (#5019)
PR #4818 (merged last week) reintroduced a config check for vocab_size that was addressed in PR #4258 (merged 2023-11-30).
Without the fix, llama2 models can't be converted. The error is:
`ValueError: The model's vocab size is set to -1 in params.json. Please update it manually. Maybe 32000?`
-rwxr-xr-x | convert.py | 2 |
1 files changed, 1 insertions, 1 deletions
@@ -348,7 +348,7 @@ class Params: f_rope_freq_base = 1e6 return Params( - n_vocab=config.get("vocab_size", model["tok_embeddings.weight"].shape[0]), + n_vocab=model["tok_embeddings.weight"].shape[0], n_embd=config["dim"], n_layer=config["n_layers"], n_ctx=n_ctx, |