|
1 |
| - |
2 |
| - |
3 |
| -# AutoGGUF - automated GGUF model quantizer |
4 |
| - |
5 |
| -[](https://github.com/ggerganov/llama.cpp) |
6 |
| - |
7 |
| - |
8 |
| - |
9 |
| - |
10 |
| - |
11 |
| - |
12 |
| - |
13 |
| - |
14 |
| -AutoGGUF provides a graphical user interface for quantizing GGUF models using the llama.cpp library. It allows users to download different versions of llama.cpp, manage multiple backends, and perform quantization tasks with various options. |
15 |
| - |
16 |
| -## Features |
17 |
| - |
18 |
| -- Download and manage llama.cpp backends |
19 |
| -- Select and quantize GGUF models |
20 |
| -- Configure quantization parameters |
21 |
| -- Monitor system resources during quantization |
22 |
| - |
23 |
| -## Usage |
24 |
| - |
25 |
| -### Cross-platform |
26 |
| -1. Install dependencies: |
27 |
| - ``` |
28 |
| - pip install -r requirements.txt |
29 |
| - ``` |
30 |
| - or |
31 |
| - ``` |
32 |
| - pip install PyQt6 requests psutil shutil |
33 |
| - ``` |
34 |
| -2. Run the application: |
35 |
| - ``` |
36 |
| - python src/main.py |
37 |
| - ``` |
38 |
| - or use the `run.bat` script. |
39 |
| - |
40 |
| -### Windows |
41 |
| -1. Download the latest release |
42 |
| -2. Extract all files to a folder |
43 |
| -3. Run `AutoGGUF.exe` |
44 |
| - |
45 |
| -## Building |
46 |
| - |
47 |
| -### Cross-platform |
48 |
| -```bash |
49 |
| -cd src |
50 |
| -pip install -U pyinstaller |
51 |
| -pyinstaller main.py --onefile |
52 |
| -cd dist/main |
53 |
| -./main |
54 |
| -``` |
55 |
| - |
56 |
| -### Windows |
57 |
| -```bash |
58 |
| -build RELEASE/DEV |
59 |
| -``` |
60 |
| -Find the executable in `build/<type>/dist/AutoGGUF.exe`. |
61 |
| - |
62 |
| -## Dependencies |
63 |
| - |
64 |
| -- PyQt6 |
65 |
| -- requests |
66 |
| -- psutil |
67 |
| -- shutil |
68 |
| -- OpenSSL |
69 |
| - |
70 |
| -## Localizations |
71 |
| - |
72 |
| -View the list of supported languages at [AutoGGUF/wiki/Installation#configuration](https://github.com/leafspark/AutoGGUF/wiki/Installation#configuration) (LLM translated, except for English). |
73 |
| - |
74 |
| -To use a specific language, set the `AUTOGGUF_LANGUAGE` environment variable to one of the listed language codes. |
75 |
| - |
76 |
| -## Known Issues |
77 |
| - |
78 |
| -- Saving preset while quantizing causes UI thread crash (planned fix: remove this feature) |
79 |
| -- Cannot delete task while processing (planned fix: disallow deletion before cancelling or cancel automatically) |
80 |
| -- Base Model text still shows when GGML is selected as LoRA type (fix: include text in show/hide Qt layout) |
81 |
| - |
82 |
| -## Planned Features |
83 |
| - |
84 |
| -- Actual progress bar tracking |
85 |
| -- Download safetensors from HF and convert to unquantized GGUF |
86 |
| -- Perplexity testing |
87 |
| -- Managing shards (coming in the next release) |
88 |
| -- Time estimation for quantization |
89 |
| -- Dynamic values for KV cache (coming in the next release) |
90 |
| -- Ability to select and start multiple quants at once (saved in presets, coming in the next release) |
91 |
| - |
92 |
| -## Troubleshooting |
93 |
| - |
94 |
| -- SSL module cannot be found error: Install OpenSSL or run from source using `python src/main.py` with the `run.bat` script (`pip install requests`) |
95 |
| - |
96 |
| -## Contributing |
97 |
| - |
98 |
| -Fork the repo, make your changes, and ensure you have the latest commits when merging. Include a changelog of new features in your pull request description. |
99 |
| - |
100 |
| -## User Interface |
101 |
| - |
102 |
| - |
103 |
| - |
104 |
| -## Stargazers |
105 |
| - |
106 |
| -[](https://star-history.com/#leafspark/AutoGGUF&Date) |
| 1 | + |
| 2 | + |
| 3 | +# AutoGGUF - automated GGUF model quantizer |
| 4 | + |
| 5 | +[](https://github.com/ggerganov/llama.cpp) |
| 6 | + |
| 7 | + |
| 8 | + |
| 9 | + |
| 10 | + |
| 11 | + |
| 12 | + |
| 13 | + |
| 14 | +AutoGGUF provides a graphical user interface for quantizing GGUF models using the llama.cpp library. It allows users to download different versions of llama.cpp, manage multiple backends, and perform quantization tasks with various options. |
| 15 | + |
| 16 | +## Features |
| 17 | + |
| 18 | +- Download and manage llama.cpp backends |
| 19 | +- Select and quantize GGUF models |
| 20 | +- Configure quantization parameters |
| 21 | +- Monitor system resources during quantization |
| 22 | + |
| 23 | +## Usage |
| 24 | + |
| 25 | +### Cross-platform |
| 26 | +1. Install dependencies: |
| 27 | + ``` |
| 28 | + pip install -r requirements.txt |
| 29 | + ``` |
| 30 | + or |
| 31 | + ``` |
| 32 | + pip install PyQt6 requests psutil shutil |
| 33 | + ``` |
| 34 | +2. Run the application: |
| 35 | + ``` |
| 36 | + python src/main.py |
| 37 | + ``` |
| 38 | + or use the `run.bat` script. |
| 39 | + |
| 40 | +### Windows |
| 41 | +1. Download the latest release |
| 42 | +2. Extract all files to a folder |
| 43 | +3. Run `AutoGGUF.exe` |
| 44 | + |
| 45 | +## Building |
| 46 | + |
| 47 | +### Cross-platform |
| 48 | +```bash |
| 49 | +cd src |
| 50 | +pip install -U pyinstaller |
| 51 | +pyinstaller main.py --onefile |
| 52 | +cd dist/main |
| 53 | +./main |
| 54 | +``` |
| 55 | + |
| 56 | +### Windows |
| 57 | +```bash |
| 58 | +build RELEASE/DEV |
| 59 | +``` |
| 60 | +Find the executable in `build/<type>/dist/AutoGGUF.exe`. |
| 61 | + |
| 62 | +## Dependencies |
| 63 | + |
| 64 | +- PyQt6 |
| 65 | +- requests |
| 66 | +- psutil |
| 67 | +- shutil |
| 68 | +- OpenSSL |
| 69 | + |
| 70 | +## Localizations |
| 71 | + |
| 72 | +View the list of supported languages at [AutoGGUF/wiki/Installation#configuration](https://github.com/leafspark/AutoGGUF/wiki/Installation#configuration) (LLM translated, except for English). |
| 73 | + |
| 74 | +To use a specific language, set the `AUTOGGUF_LANGUAGE` environment variable to one of the listed language codes. |
| 75 | + |
| 76 | +## Known Issues |
| 77 | + |
| 78 | +- Saving preset while quantizing causes UI thread crash (planned fix: remove this feature) |
| 79 | +- Cannot delete task while processing (planned fix: disallow deletion before cancelling or cancel automatically) |
| 80 | +- Base Model text still shows when GGML is selected as LoRA type (fix: include text in show/hide Qt layout) |
| 81 | + |
| 82 | +## Planned Features |
| 83 | + |
| 84 | +- Actual progress bar tracking |
| 85 | +- Download safetensors from HF and convert to unquantized GGUF |
| 86 | +- Perplexity testing |
| 87 | +- Managing shards (coming in the next release) |
| 88 | +- Time estimation for quantization |
| 89 | +- Dynamic values for KV cache (coming in the next release) |
| 90 | +- Ability to select and start multiple quants at once (saved in presets, coming in the next release) |
| 91 | + |
| 92 | +## Troubleshooting |
| 93 | + |
| 94 | +- SSL module cannot be found error: Install OpenSSL or run from source using `python src/main.py` with the `run.bat` script (`pip install requests`) |
| 95 | + |
| 96 | +## Contributing |
| 97 | + |
| 98 | +Fork the repo, make your changes, and ensure you have the latest commits when merging. Include a changelog of new features in your pull request description. |
| 99 | + |
| 100 | +## User Interface |
| 101 | + |
| 102 | + |
| 103 | + |
| 104 | +## Stargazers |
| 105 | + |
| 106 | +[](https://star-history.com/#leafspark/AutoGGUF&Date) |
0 commit comments