# Changelog All notable changes to this project will be documented in this file. ## [1.6.0] - 2024-08-08 ### Changed - Resolve licensing issues by using PySide6 ### Added - Add GPU monitoring support for NVIDIA GPUs ## [1.5.1] - 2024-08-08 ### Changed - Refactor localizations to use them in HF conversion area - Rename FAILED_LOAD_PRESET to FAILED_TO_LOAD_PRESET localization key ### Removed - Remove Save Preset context menu action ### Added - Support loading *.gguf file types ## [1.5.0] - 2024-08-06 ### Changed - Refactor localizations to use them in HF conversion area - Organize localizations ### Added - Add sha256 and PGP signatures (same as commit ones) - Add HuggingFace to GGUF conversion support ### Fixed - Fix scaling on low resolution screens, interface now scrolls ## [1.4.3] - 2024-08-05 ### Changed - Updated src file in release to be Black formatted - Modifying the quantize_model function to process all selected types - Updating preset saving and loading to handle multiple quantization types - Use ERROR and IN_PROGRESS constants from localizations in QuantizationThread - Minor repository changes ### Added - Added model sharding management support - Allow multiple quantization types to be selected and started simultaneously ## [1.4.2] - 2024-08-04 ### Fixed - Resolves bug where Base Model text was shown even when GGML type was selected - Improved alignment ### Changed - Minor repository changes ## [1.4.1] - 2024-08-04 ### Added - Dynamic KV Overrides (see wiki: AutoGGUF/wiki/Dynamic-KV-Overrides) - Quantization commands are now printed and logged ## [1.4.0] - 2024-08-04 ### Added - LoRA Conversion: - New section for converting HuggingFace PEFT LoRA adapters to GGML/GGUF - Output type selection (GGML or GGUF) - Base model selection for GGUF output - LoRA adapter list with individual scaling factors - Export LoRA section for merging adapters into base model - UI Improvements: - Updated task names in task list - IMatrix generation check - Larger window size - Added exe favicon - Localization: - French and Simplified Chinese support for LoRA and "Refresh Models" strings - Code and Build: - Code organization improvements - Added build script - .gitignore file - Misc: - Currently includes src folder with conversion tools - No console window popup ## [1.3.1] - 2024-08-04 ### Added - AUTOGGUF_CHECK_BACKEND environment variable to disable backend check on start ### Changed - --onefile build with PyInstaller, _internal directory is no longer required ## [1.3.0] - 2024-08-03 ### Added - Support for new llama-imatrix parameters: - Context size (--ctx-size) input - Threads (--threads) control - New parameters to IMatrix section layout - Slider-spinbox combination for thread count selection - QSpinBox for output frequency input (1-100 range with percentage suffix) ### Changed - Converted context size input to a QSpinBox - Updated generate_imatrix() method to use new UI element values - Improved error handling in preset loading - Enhanced localization support for new UI elements ### Fixed - Error when loading presets containing KV overrides ### Removed - Duplicated functions ## [1.2.1] - 2024-08-03 ### Added - Refresh Models button - Linux build (built on Ubuntu 24.04 LTS) ### Fixed - iostream llama.cpp issue, quantized_models directory created on launch ## [1.2.0] - 2024-08-03 ### Added - More robust logging (find logs at latest_.log in logs folder) - Localizations with support for 28 languages (machine translated using Gemini Experimental 0801) ## [1.1.0] - 2024-08-03 ### Added - Dynamic KV override functionality - Improved CUDA checking ability and extraction to the backend folder - Scrollable area for KV overrides with add/delete capabilities ### Changed - Enhanced visibility and usability of Output Tensor Type and Token Embedding Type options - Refactored code for better modularity and reduced circular dependencies ### Fixed - Behavior of Output Tensor Type and Token Embedding Type dropdown menus - Various minor UI inconsistencies ## [1.0.1] - 2024-08-02 ### Added - Windows binary (created using PyInstaller) ### Fixed - Issue where quantization errored with "AutoGGUF does not have x attribute" ## [1.0.0] - 2024-08-02 ### Added - Initial release - GUI interface for automated GGUF model quantization - System resource monitoring (RAM and CPU usage) - Llama.cpp backend selection and management - Automatic download of llama.cpp releases from GitHub - Model selection from local directory - Comprehensive quantization options - Task list for managing multiple quantization jobs - Real-time log viewing for quantization tasks - IMatrix generation feature with customizable settings - GPU offload settings for IMatrix generation - Context menu for task management - Detailed model information dialog - Error handling and user notifications - Confirmation dialogs for task deletion and application exit