You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
When I pass the --cml or --chatml commandline options to llamafile, it exits with the error message:
error: unknown argument: -cml
OR
error: unknown argument: --chatml
Also, the reason why I am attempting to use this commandline option is because I am running with the model dolphin-2.9-llama3-8b-Q5_K_M.gguf and every line of LLM output has the suffix <|im_end|> appended to it, which is annoying. I wondered if enabling chatml mode would filter it out.
Example:
User: This is a test message
Llama: Hello there! How can I assist you today?
<|im_end|>
User: Why are you including that token at the end?
Llama: My apologies for that. It seems my response was unintentionally influenced by my underlying system message format. Let me correct myself and provide an accurate answer to your question.
What is the reason behind including a specific token in a conversation between users and chatbots?
<|im_end|>
The text was updated successfully, but these errors were encountered:
I'm running llamafile 0.8.1
When I pass the --cml or --chatml commandline options to llamafile, it exits with the error message:
error: unknown argument: -cml
OR
error: unknown argument: --chatml
Also, the reason why I am attempting to use this commandline option is because I am running with the model dolphin-2.9-llama3-8b-Q5_K_M.gguf and every line of LLM output has the suffix <|im_end|> appended to it, which is annoying. I wondered if enabling chatml mode would filter it out.
Example:
The text was updated successfully, but these errors were encountered: