You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
Traceback (most recent call last):
File "/home/local/vanroy/llama.cpp/convert-hf-to-gguf.py", line 3001, in <module>
main()
File "/home/local/vanroy/llama.cpp/convert-hf-to-gguf.py", line 2988, in main
model_instance.set_vocab()
File "/home/local/vanroy/llama.cpp/convert-hf-to-gguf.py", line 75, in set_vocab
self._set_vocab_gpt2()
File "/home/local/vanroy/llama.cpp/convert-hf-to-gguf.py", line 331, in _set_vocab_gpt2
tokens, toktypes, tokpre = self.get_vocab_base()
File "/home/local/vanroy/llama.cpp/convert-hf-to-gguf.py", line 242, in get_vocab_base
tokpre = self.get_vocab_base_pre(tokenizer)
File "/home/local/vanroy/llama.cpp/convert-hf-to-gguf.py", line 323, in get_vocab_base_pre
raise NotImplementedError("BPE pre-tokenizer was not recognized - update get_vocab_base_pre()")
NotImplementedError: BPE pre-tokenizer was not recognized - update get_vocab_base_pre()
Prerequisites
Please answer the following questions for yourself before submitting an issue.
Feature Description
Provide support for phi-2. Running the following yields an error:
Error:
Phi-2 uses
CodeGenTokenizer
, which is a BPE Tokenizer.I'm not sure if it is as easy as adding the following line here?
Edit tried that, this is the generated hash:
The text was updated successfully, but these errors were encountered: