AutoGGUF/CHANGELOG.md

230 lines
6.4 KiB
Markdown

# Changelog
## [v1.7.0] - 2024-08-16
### Added
- Menu bar with Close and About options
- Program version in localizations.py
- Support for 32-bit builds
- Added dependency audit
- Implemented radon, dependabot, and pre-commit workflows
### Changed
- Updated torch requirement from `~=1.13.1` to `~=2.4.0`
- Updated psutil requirement from `~=5.9.8` to `~=6.0.0`
- Refactored functions out of AutoGGUF.py and moved to ui_update.py
- Changed filenames to follow PEP 8 conventions
- Disabled .md and .txt CodeQL analysis
### Fixed
- Optimized imports in AutoGGUF.py
- Updated README with new version and styled screenshot
- Fixed image blur in documentation
## [v1.6.2] - 2024-08-15
### Added
- Server functionality with new endpoints:
- `/v1/backends`: Lists all backends and their paths
- `/v1/health`: Heartbeat endpoint
- `/v1/tasks`: Provides current task info (name, status, progress, log file)
- `/v1/models`: Retrieves model details (name, type, path, shard status)
- Environment variable support for server configuration:
- `AUTOGGUF_SERVER`: Enable/disable server (true/false)
- `AUTOGGUF_SERVER_PORT`: Set server port (integer)
### Changed
- Updated AutoGGUF docstrings
- Refactored build scripts
### Fixed
- Set GGML types to lowercase in command builder
## [v1.6.1] - 2024-08-12
### Added
- Optimized build scripts
- Nuitka for building
### Changed
- Updated .gitignore
### Fixed
- Bug where deletion while a task is running crashes the program
### Notes
- Fast build: Higher unzipped size (97MB), smaller download (38MB)
- Standard build: Created with PyInstaller, medium download and unzipped size (50MB), potentially slower
## [1.6.0] - 2024-08-08
### Changed
- Resolve licensing issues by using PySide6
### Added
- Add GPU monitoring support for NVIDIA GPUs
## [1.5.1] - 2024-08-08
### Changed
- Refactor localizations to use them in HF conversion area
- Rename FAILED_LOAD_PRESET to FAILED_TO_LOAD_PRESET localization key
### Removed
- Remove Save Preset context menu action
### Added
- Support loading *.gguf file types
## [1.5.0] - 2024-08-06
### Changed
- Refactor localizations to use them in HF conversion area
- Organize localizations
### Added
- Add sha256 and PGP signatures (same as commit ones)
- Add HuggingFace to GGUF conversion support
### Fixed
- Fix scaling on low resolution screens, interface now scrolls
## [1.4.3] - 2024-08-05
### Changed
- Updated src file in release to be Black formatted
- Modifying the quantize_model function to process all selected types
- Updating preset saving and loading to handle multiple quantization types
- Use ERROR and IN_PROGRESS constants from localizations in QuantizationThread
- Minor repository changes
### Added
- Added model sharding management support
- Allow multiple quantization types to be selected and started simultaneously
## [1.4.2] - 2024-08-04
### Fixed
- Resolves bug where Base Model text was shown even when GGML type was selected
- Improved alignment
### Changed
- Minor repository changes
## [1.4.1] - 2024-08-04
### Added
- Dynamic KV Overrides (see wiki: AutoGGUF/wiki/Dynamic-KV-Overrides)
- Quantization commands are now printed and logged
## [1.4.0] - 2024-08-04
### Added
- LoRA Conversion:
- New section for converting HuggingFace PEFT LoRA adapters to GGML/GGUF
- Output type selection (GGML or GGUF)
- Base model selection for GGUF output
- LoRA adapter list with individual scaling factors
- Export LoRA section for merging adapters into base model
- UI Improvements:
- Updated task names in task list
- IMatrix generation check
- Larger window size
- Added exe favicon
- Localization:
- French and Simplified Chinese support for LoRA and "Refresh Models" strings
- Code and Build:
- Code organization improvements
- Added build script
- .gitignore file
- Misc:
- Currently includes src folder with conversion tools
- No console window popup
## [1.3.1] - 2024-08-04
### Added
- AUTOGGUF_CHECK_BACKEND environment variable to disable backend check on start
### Changed
- --onefile build with PyInstaller, _internal directory is no longer required
## [1.3.0] - 2024-08-03
### Added
- Support for new llama-imatrix parameters:
- Context size (--ctx-size) input
- Threads (--threads) control
- New parameters to IMatrix section layout
- Slider-spinbox combination for thread count selection
- QSpinBox for output frequency input (1-100 range with percentage suffix)
### Changed
- Converted context size input to a QSpinBox
- Updated generate_imatrix() method to use new UI element values
- Improved error handling in preset loading
- Enhanced localization support for new UI elements
### Fixed
- Error when loading presets containing KV overrides
### Removed
- Duplicated functions
## [1.2.1] - 2024-08-03
### Added
- Refresh Models button
- Linux build (built on Ubuntu 24.04 LTS)
### Fixed
- iostream llama.cpp issue, quantized_models directory created on launch
## [1.2.0] - 2024-08-03
### Added
- More robust logging (find logs at latest_<timestamp>.log in logs folder)
- Localizations with support for 28 languages (machine translated using Gemini Experimental 0801)
## [1.1.0] - 2024-08-03
### Added
- Dynamic KV override functionality
- Improved CUDA checking ability and extraction to the backend folder
- Scrollable area for KV overrides with add/delete capabilities
### Changed
- Enhanced visibility and usability of Output Tensor Type and Token Embedding Type options
- Refactored code for better modularity and reduced circular dependencies
### Fixed
- Behavior of Output Tensor Type and Token Embedding Type dropdown menus
- Various minor UI inconsistencies
## [1.0.1] - 2024-08-02
### Added
- Windows binary (created using PyInstaller)
### Fixed
- Issue where quantization errored with "AutoGGUF does not have x attribute"
## [1.0.0] - 2024-08-02
### Added
- Initial release
- GUI interface for automated GGUF model quantization
- System resource monitoring (RAM and CPU usage)
- Llama.cpp backend selection and management
- Automatic download of llama.cpp releases from GitHub
- Model selection from local directory
- Comprehensive quantization options
- Task list for managing multiple quantization jobs
- Real-time log viewing for quantization tasks
- IMatrix generation feature with customizable settings
- GPU offload settings for IMatrix generation
- Context menu for task management
- Detailed model information dialog
- Error handling and user notifications
- Confirmation dialogs for task deletion and application exit