Closed
Description
What happened?
As per discussions:
It seems to be impossible to chat with llama3 8b properly. I have not tested this on 70b models but even in the server UI the model just starts making notes to itself and output garbage / training data as to how it should converse instead of actually conversing. Has something happened to the --chat-template chatml parameter? Even when the CLI is set to output special tokens, I do not see the ChatML tokens coming out.
Name and Version
version: 3158 (5239925)
What operating system are you seeing the problem on?
Linux
Relevant log output
No response