Skip to content

[feat]: build llama-cpp-python instead of using the whl #182

New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Open
kyteinsky opened this issue May 15, 2025 · 2 comments
Open

[feat]: build llama-cpp-python instead of using the whl #182

kyteinsky opened this issue May 15, 2025 · 2 comments
Labels
enhancement New feature or request

Comments

@kyteinsky
Copy link
Contributor

Describe the feature you'd like to request

Instead of using the whl file from https://github.com/abetlen/llama-cpp-python/releases, build the package in the docker container. Newer versions of llama-cpp-python haven't had pre-built releases recently which would make building the only way to have them officially.

This will also solve #126 and #178 (comment) (using app_api's env declaration for CMAKE_ARGS).

Some related links:

@kyteinsky kyteinsky added the enhancement New feature or request label May 15, 2025
@marcelklehr
Copy link
Member

I think we should limit the configurations we support. Not because I want to exclude people, but because more options gives more surface area for failure, which we have to debug and solve. We can try to support non-AVX CPUs as well as ARM, but I'd not go so far as to just invite arbitrary CMAKE_ARGS values.

@kyteinsky
Copy link
Contributor Author

yeah fair enough. Building llama-cpp-python should alone be enough to address these.

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
enhancement New feature or request
Projects
None yet
Development

No branches or pull requests

2 participants