-
Notifications
You must be signed in to change notification settings - Fork 11.3k
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Unable to convert Smaug 72B #5807
Comments
llama.cpp release b2291 btw. |
I think it works by changing this line:
|
Thanks a lot, but that just fails further down the line: Error: Missing Smaug-72B-v0.1/tokenizer.model I notice quite a lot of models on huggingface were apparently convertible a few weeks/months ago but no longer are (e..g TheBloke has GGUFs, but when I try converting with current versions of llama to make imatrix quants, they fail in lots of different ways). Is this considered a regression and should issues be created for those? |
Ah yes, I think I recreated it by loading the model using HF transformers and using |
It seems most of the llama-2 derived models seem to have this vocabulary mismatch problem, and #5821 does not seem to help. |
Any updates/workarounds on this? I just tried to convert Smaug 72B and I'm getting a |
@schmorp, @dranger003 & @christiandaley I am using the latest version of https://huggingface.co/abacusai/Smaug-72B-v0.1 and the llama.cpp release b2405. Solution:
python convert.py --vocab-type "bpe" --pad-vocab --outfile ./models/Smaug-72B-v0.1.gguf /path/to/repository Server Log:
Open Question: Is the warning something to worry about?
|
I thought I'd have tried that, but maybe I haven't. Thanks a lot for this tip! As for the special token deifnition warning, I had this for a few other models, and they seemed to work, but ymmv. |
Works, so it's a user error. Sorry for the noise. |
When converting with the command from @countzero, the resulting model crashes main/imatrix with: terminate called after throwing an instance of 'std::out_of_range' Sorry, @christiandaley, you actually reported this a week ago and I overlooked it. |
This issue was closed because it has been inactive for 14 days since being marked as stale. |
I am unable to convert https://huggingface.co/abacusai/Smaug-72B-v0.1 (and others) to GGUF with either convert.py or convert-hf-to-gguf.py.
With the former, I get:
RuntimeError: Internal: ./src/sentencepiece_processor.cc(1101) [model_proto->ParseFromArray(serialized.data(), serialized.size())]
"internal" feels like a bug. When I add --vocab-type hfft (and then --pad-vocab because it tells me to), I get a nonfunctional model:
llm_load_vocab: SPM vocabulary, but newline token not found: unordered_map::at! Using special_pad_id instead.llm_load_vocab: mismatch in special tokens definition ( 421/152064 vs 214/152064 ).
and convert-hf-to-gguf.py does not support "LlamaForCausalLM".
The text was updated successfully, but these errors were encountered: