-
I recently came across with ChatDoctorrepository, it looks very interesting but I haven't been able to generate working ggml templates so far to be able to run via llama.cpp (master). Got this error while trying to export to .pth: There was no error while running the convertion from .pth to .ggml or while running the quantization to get ggml-model-q4_0.bin, but I can't execute the ggml-model-q4_0.bin model from llama.cpp. Got the following message:
Don't know if this a issue. I guess I'm doing something wrong while exporting to pth. Until now I was following and trying to adapt the following process, so don't know if I should open a issue or even if such model is compatible with llama.cpp at moment. Any help will be appreciated. |
Beta Was this translation helpful? Give feedback.
Replies: 5 comments 4 replies
-
I remember vaguely that the whole LLaMATokenizer vs LlamaTokenizer was because you need to update your |
Beta Was this translation helpful? Give feedback.
-
try also editing the
change |
Beta Was this translation helpful? Give feedback.
-
@tomsnunes Where have you downloaded the model files? |
Beta Was this translation helpful? Give feedback.
-
Where to find ggml compatible datasets/how to convert them? |
Beta Was this translation helpful? Give feedback.
I remember vaguely that the whole LLaMATokenizer vs LlamaTokenizer was because you need to update your
transformers
python library to the latest version.