|
| 1 | +# Changelog |
| 2 | + |
| 3 | +All notable changes to this project will be documented in this file. |
| 4 | + |
| 5 | +## [1.4.1] - 2024-08-04 |
| 6 | + |
| 7 | +### Added |
| 8 | +- Dynamic KV Overrides (see wiki: AutoGGUF/wiki/Dynamic-KV-Overrides) |
| 9 | +- Quantization commands are now printed and logged |
| 10 | + |
| 11 | +## [1.4.0] - 2024-08-04 |
| 12 | + |
| 13 | +### Added |
| 14 | +- LoRA Conversion: |
| 15 | + - New section for converting HuggingFace PEFT LoRA adapters to GGML/GGUF |
| 16 | + - Output type selection (GGML or GGUF) |
| 17 | + - Base model selection for GGUF output |
| 18 | + - LoRA adapter list with individual scaling factors |
| 19 | + - Export LoRA section for merging adapters into base model |
| 20 | +- UI Improvements: |
| 21 | + - Updated task names in task list |
| 22 | + - IMatrix generation check |
| 23 | + - Larger window size |
| 24 | + - Added exe favicon |
| 25 | +- Localization: |
| 26 | + - French and Simplified Chinese support for LoRA and "Refresh Models" strings |
| 27 | +- Code and Build: |
| 28 | + - Code organization improvements |
| 29 | + - Added build script |
| 30 | + - .gitignore file |
| 31 | +- Misc: |
| 32 | + - Currently includes src folder with conversion tools |
| 33 | + - No console window popup |
| 34 | + |
| 35 | +## [1.3.1] - 2024-08-04 |
| 36 | + |
| 37 | +### Added |
| 38 | +- AUTOGGUF_CHECK_BACKEND environment variable to disable backend check on start |
| 39 | + |
| 40 | +### Changed |
| 41 | +- --onefile build with PyInstaller, _internal directory is no longer required |
| 42 | + |
| 43 | +## [1.3.0] - 2024-08-03 |
| 44 | + |
| 45 | +### Added |
| 46 | +- Support for new llama-imatrix parameters: |
| 47 | + - Context size (--ctx-size) input |
| 48 | + - Threads (--threads) control |
| 49 | +- New parameters to IMatrix section layout |
| 50 | +- Slider-spinbox combination for thread count selection |
| 51 | +- QSpinBox for output frequency input (1-100 range with percentage suffix) |
| 52 | + |
| 53 | +### Changed |
| 54 | +- Converted context size input to a QSpinBox |
| 55 | +- Updated generate_imatrix() method to use new UI element values |
| 56 | +- Improved error handling in preset loading |
| 57 | +- Enhanced localization support for new UI elements |
| 58 | + |
| 59 | +### Fixed |
| 60 | +- Error when loading presets containing KV overrides |
| 61 | + |
| 62 | +### Removed |
| 63 | +- Duplicated functions |
| 64 | + |
| 65 | +## [1.2.1] - 2024-08-03 |
| 66 | + |
| 67 | +### Added |
| 68 | +- Refresh Models button |
| 69 | +- Linux build (built on Ubuntu 24.04 LTS) |
| 70 | + |
| 71 | +### Fixed |
| 72 | +- iostream llama.cpp issue, quantized_models directory created on launch |
| 73 | + |
| 74 | +## [1.2.0] - 2024-08-03 |
| 75 | + |
| 76 | +### Added |
| 77 | +- More robust logging (find logs at latest_<timestamp>.log in logs folder) |
| 78 | +- Localizations with support for 28 languages (machine translated using Gemini Experimental 0801) |
| 79 | + |
| 80 | +## [1.1.0] - 2024-08-03 |
| 81 | + |
| 82 | +### Added |
| 83 | +- Dynamic KV override functionality |
| 84 | +- Improved CUDA checking ability and extraction to the backend folder |
| 85 | +- Scrollable area for KV overrides with add/delete capabilities |
| 86 | + |
| 87 | +### Changed |
| 88 | +- Enhanced visibility and usability of Output Tensor Type and Token Embedding Type options |
| 89 | +- Refactored code for better modularity and reduced circular dependencies |
| 90 | + |
| 91 | +### Fixed |
| 92 | +- Behavior of Output Tensor Type and Token Embedding Type dropdown menus |
| 93 | +- Various minor UI inconsistencies |
| 94 | + |
| 95 | +## [1.0.1] - 2024-08-02 |
| 96 | + |
| 97 | +### Added |
| 98 | +- Windows binary (created using PyInstaller) |
| 99 | + |
| 100 | +### Fixed |
| 101 | +- Issue where quantization errored with "AutoGGUF does not have x attribute" |
| 102 | + |
| 103 | +## [1.0.0] - 2024-08-02 |
| 104 | + |
| 105 | +### Added |
| 106 | +- Initial release |
| 107 | +- GUI interface for automated GGUF model quantization |
| 108 | +- System resource monitoring (RAM and CPU usage) |
| 109 | +- Llama.cpp backend selection and management |
| 110 | +- Automatic download of llama.cpp releases from GitHub |
| 111 | +- Model selection from local directory |
| 112 | +- Comprehensive quantization options |
| 113 | +- Task list for managing multiple quantization jobs |
| 114 | +- Real-time log viewing for quantization tasks |
| 115 | +- IMatrix generation feature with customizable settings |
| 116 | +- GPU offload settings for IMatrix generation |
| 117 | +- Context menu for task management |
| 118 | +- Detailed model information dialog |
| 119 | +- Error handling and user notifications |
| 120 | +- Confirmation dialogs for task deletion and application exit |
0 commit comments