Skip to content

Prebuilt CUDA wheels not working #1822

@mjwweb

Description

@mjwweb

There are multiple issues with the CUDA wheels:

  1. The cu125 repository returns 404:

    $ curl -I https://abetlen.github.io/llama-cpp-python/whl/cu125/
    HTTP/2 404
  2. While cu124 exists, pip fails to find wheels using --extra-index-url:

    pip install llama-cpp-python --extra-index-url https://abetlen.github.io/llama-cpp-python/whl/cu124
  3. Even direct wheel installation fails due to missing CUDA dependencies:

    pip install https://github.com/abetlen/llama-cpp-python/releases/download/v0.2.90-cu124/llama_cpp_python-0.2.90-cp312-cp312-linux_x86_64.whl

    Results in:

    RuntimeError: Failed to load shared library '.../libllama.so': libcudart.so.12: cannot open shared object file: No such file or directory
    

This setup was working a few weeks ago without requiring manual CUDA installation.

Environment:

  • Python 3.12
  • pip 24.2
  • WSL2 Ubuntu

Metadata

Metadata

Assignees

No one assigned

    Labels

    No labels
    No labels

    Projects

    No projects

    Milestone

    No milestone

    Relationships

    None yet

    Development

    No branches or pull requests

    Issue actions