-
Notifications
You must be signed in to change notification settings - Fork 961
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
FileNotFoundError: Shared library with base name 'llama' not found #568
Comments
The repo is cloned recursively and I am able to go into the vendor directory and compile llama_cpp and run it.
and that works fine. Going back to llama-cpp-python and trying to load the library didn't work. I even tried:
With a shared libs option turned on, but no dice. |
Ok weird. If I create a from llama_cpp import Llama and launch it:
It works perfectly. So it's only the interactive Python that is having a problem. Ok... |
Hi, I had similar issues. As it turns out, the problem was, that the "wrong" llama_cpp.py was used to perform the import. Thats a problem because llama_cpp.py::_load_shared_library() is using I am not sure, but I think instead of using |
You're exactly right, when I move out of the project directory I can suddenly do @abetlen Please take a look at this issue |
Thank you both - I had the same experience. Within the |
Sounds like something needs to be modified in the code ... |
Thanks to all that posted. This bug was driving me crazy. |
I am seeing a similar error when trying to start llama-cpp-pyton container for llama-gpt: getumbrel/llama-gpt#144 . Any idea if it is caused by the same problem ? The stacktrace looks similar... |
Thank you for posting the workaround! |
FileNotFoundError: Shared library with base name 'llama' not found,please tell me how to deal with ?thanks!!! |
Prerequisites
Please answer the following questions for yourself before submitting an issue.
Expected Behavior
I'm following the instructions on the README. llama_cpp is buildable on my machine with cuBLAS support (libraries and paths are correct).
The installation seems to go well:
I expected to be able to import the library but that doesn't work.
Current Behavior
Environment and Context
$ lscpu
$ uname -a
Steps to Reproduce
Please provide detailed steps for reproducing the issue. We are not sitting in front of your screen, so the more detail the better.
Note: Many issues seem to be regarding functional or performance issues / differences with
llama.cpp
. In these cases we need to confirm that you're comparing against the version ofllama.cpp
that was built with your python package, and which parameters you're passing to the context.Try the following:
git clone https://github.com/abetlen/llama-cpp-python
cd llama-cpp-python
rm -rf _skbuild/
# delete any old buildspython setup.py develop
cd ./vendor/llama.cpp
cmake
llama.cpp./main
with the same arguments you previously passed to llama-cpp-python and see if you can reproduce the issue. If you can, log an issue with llama.cppI tried, and I get this:
The text was updated successfully, but these errors were encountered: