Mozilla-Ocho / llamafile

Distribute and run LLMs with a single file.

Home Page:https://llamafile.ai

Geek Repo:Geek Repo

Github PK Tool:Github PK Tool

chatml/cml commandline option rejected

vlasky opened this issue · comments

I'm running llamafile 0.8.1

When I pass the --cml or --chatml commandline options to llamafile, it exits with the error message:

error: unknown argument: -cml

OR

error: unknown argument: --chatml

Also, the reason why I am attempting to use this commandline option is because I am running with the model dolphin-2.9-llama3-8b-Q5_K_M.gguf and every line of LLM output has the suffix <|im_end|> appended to it, which is annoying. I wondered if enabling chatml mode would filter it out.

Example:

User: This is a test message

Llama: Hello there! How can I assist you today?

<|im_end|>

User: Why are you including that token at the end?

Llama: My apologies for that. It seems my response was unintentionally influenced by my underlying system message format. Let me correct myself and provide an accurate answer to your question.

What is the reason behind including a specific token in a conversation between users and chatbots?

<|im_end|>

Same here.