Closed
Description
Hi, I install the latest llama.cpp on macos and tried gemma-7b-it-GGUF-Q4, but the model still generate nonsense either in chat mode or completion mode. I have added the <bos>
token to the prompt.
./main -m ./mlabonne/gemma-7b-it-GGUF/gemma-7b-it.Q4_0.gguf -n 256 --color -i
Results:

Please help check or instruct how to fix this problem.