There are multiple issues with the CUDA wheels:
-
The cu125 repository returns 404:
$ curl -I https://abetlen.github.io/llama-cpp-python/whl/cu125/
HTTP/2 404
-
While cu124 exists, pip fails to find wheels using --extra-index-url:
pip install llama-cpp-python --extra-index-url https://abetlen.github.io/llama-cpp-python/whl/cu124
-
Even direct wheel installation fails due to missing CUDA dependencies:
pip install https://github.com/abetlen/llama-cpp-python/releases/download/v0.2.90-cu124/llama_cpp_python-0.2.90-cp312-cp312-linux_x86_64.whl
Results in:
RuntimeError: Failed to load shared library '.../libllama.so': libcudart.so.12: cannot open shared object file: No such file or directory
This setup was working a few weeks ago without requiring manual CUDA installation.
Environment:
- Python 3.12
- pip 24.2
- WSL2 Ubuntu
There are multiple issues with the CUDA wheels:
The cu125 repository returns 404:
While cu124 exists, pip fails to find wheels using --extra-index-url:
Even direct wheel installation fails due to missing CUDA dependencies:
Results in:
This setup was working a few weeks ago without requiring manual CUDA installation.
Environment: