You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
After reviewing multiple GitHub issues, forum discussions, and guides from other Python packages, I was able to successfully build and install llama-cpp-python 0.3.7 with CUDA on Windows 11. Below are the exact steps I followed, which may help others facing similar build issues.
Errors Encountered During Build
I encountered the following build errors even after installing the right build tools. The issue was resolved by updating the llama.cpp submodule.
D:\llama-cpp-python\vendor\llama.cpp\common\log.cpp(28,79): error C2039: 'system_clock': is not a member of 'std::chrono' [%localappdata%\Temp\tmp0zkelev6\build\vendor\llama.cpp\common\common.vcxproj]D:\llama-cpp-python\vendor\llama.cpp\common\log.cpp(28,79): error C3083: 'system_clock': the symbol to the left of a '::' must be a type [%localappdata%\Temp\tmp0zkelev6\build\vendor\llama.cpp\common\common.vcxproj]D:\llama-cpp-python\vendor\llama.cpp\common\log.cpp(28,93): error C2039: 'now': is not a member of 'std::chrono' [%localappdata%\Temp\tmp0zkelev6\build\vendor\llama.cpp\common\common.vcxproj]D:\llama-cpp-python\vendor\llama.cpp\common\log.cpp(28,93): error C3861: 'now': identifier not found [%localappdata%\Temp\tmp0zkelev6\build\vendor\llama.cpp\common\common.vcxproj]D:\llama-cpp-python\vendor\llama.cpp\common\log.cpp(28,25): error C2672: 'std::chrono::duration_cast': no matching overloaded function found [%localappdata%\Temp\tmp0zkelev6\build\vendor\llama.cpp\common\common.vcxproj]
Hardware Setup
OS: Windows 11
CPU: Intel i9-9900K
GPUs: RTX 3080, RTX A4000
Software Setup
Git Bash: 2.48.1
Python: 3.12.9 (likely works with 3.11 and 3.13)
CMake: 4.0.0-rc3
Visual Studio 2022: 17.13.2 with the following workloads installed -
Python development
Desktop development with C++
Linux and embedded development with C++
CUDA Toolkit: 12.8.0 (installed without modifying PhysX, Display, and HD Audio drivers, as they were already installed by the latest Studio Driver Package v572.60 that I'm using)
Installation Steps
Make sure to use Command Prompt and not PowerShell, for the following steps.
1. Clone the Repository
Navigate to the directory where you want to build the package and run:
set FORCE_CMAKE=1
set CMAKE_ARGS=-DGGML_CUDA=ON
# Optional variables to speed up the build process# set CMAKE_BUILD_PARALLEL_LEVEL=%NUMBER_OF_PROCESSORS%# set MAKEFLAGS=-j%NUMBER_OF_PROCESSORS%
These are only temporary and will be unset when you exit the current CMD session.
5. Build and Install
Run the following command to build and install the package:
Leaving the issue open for visibility. Repo maintainers can close this issue when the documentation is updated or these steps aren't relevant any more.
After reviewing multiple GitHub issues, forum discussions, and guides from other Python packages, I was able to successfully build and install
llama-cpp-python 0.3.7
with CUDA on Windows 11. Below are the exact steps I followed, which may help others facing similar build issues.Errors Encountered During Build
I encountered the following build errors even after installing the right build tools. The issue was resolved by updating the
llama.cpp
submodule.Hardware Setup
Software Setup
Installation Steps
Make sure to use Command Prompt and not PowerShell, for the following steps.
1. Clone the Repository
Navigate to the directory where you want to build the package and run:
2. Navigate to the Package Directory
cd llama-cpp-python
3. Update the
llama.cpp
Submodulegit submodule update --remote vendor\llama.cpp
4. Set Required Environment Variables
Configure the necessary environment variables:
These are only temporary and will be unset when you exit the current CMD session.
5. Build and Install
Run the following command to build and install the package:
python -m pip install . --no-cache-dir --force-reinstall --upgrade
This setup worked for me, and I hope it helps others facing similar build issues.
The text was updated successfully, but these errors were encountered: