You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
Instead of using the whl file from https://github.com/abetlen/llama-cpp-python/releases, build the package in the docker container. Newer versions of llama-cpp-python haven't had pre-built releases recently which would make building the only way to have them officially.
This will also solve #126 and #178 (comment) (using app_api's env declaration for CMAKE_ARGS).
I think we should limit the configurations we support. Not because I want to exclude people, but because more options gives more surface area for failure, which we have to debug and solve. We can try to support non-AVX CPUs as well as ARM, but I'd not go so far as to just invite arbitrary CMAKE_ARGS values.
Describe the feature you'd like to request
Instead of using the whl file from https://github.com/abetlen/llama-cpp-python/releases, build the package in the docker container. Newer versions of llama-cpp-python haven't had pre-built releases recently which would make building the only way to have them officially.
This will also solve #126 and #178 (comment) (using app_api's env declaration for
CMAKE_ARGS
).Some related links:
-DGGML_CPU_ALL_VARIANTS=ON
in ggml : add predefined list of CPU backend variants to build ggml-org/llama.cpp#10626The text was updated successfully, but these errors were encountered: