Skip to content

Inconsistent Responses with LLaMA 3.1 and LLaMA 3 Models – Prompt or Config Issue? #9191

Answered by ggerganov
dgbaenar asked this question in Q&A
Discussion options

You must be logged in to vote

Meta-Llama-3-8B-F16.gguf looks like a base model, so it won't understand the prompt format that you are using. Try to use an instruction tuned variant like Meta-Llama-3-8B-Instruct-F16.gguf. Also, change -c to a power of 2: e.g. -c 2048

Replies: 1 comment 1 reply

Comment options

You must be logged in to vote
1 reply
@dgbaenar
Comment options

Answer selected by dgbaenar
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Category
Q&A
Labels
None yet
2 participants