mirror of https://github.com/leafspark/AutoGGUF
|
||
---|---|---|
src | ||
LICENSE | ||
README.md | ||
requirements.txt | ||
run.bat |
README.md
AutoGGUF - automated GGUF model quantizer
This application provides a graphical user interface for quantizing GGUF models using the llama.cpp library. It allows users to download different versions of llama.cpp, manage multiple backends, and perform quantization tasks with various options.
Main features:
- Download and manage llama.cpp backends
- Select and quantize GGUF models
- Configure quantization parameters
- Monitor system resources during quantization
Usage:
Cross platform:
- Install dependencies, either using the
requirements.txt
file orpip install PyQt6 requests psutil
. - Run the
run.bat
script to start the application, or run the commandpython src/main.py
.
Windows:
- Download latest release, extract all to folder and run
AutoGGUF.exe
- Enjoy!
Building:
cd src
pip install -U pyinstaller
pyinstaller main.py
cd dist/main
main
Dependencies:
- PyQt6
- requests
- psutil
Issues:
- Actual progress bar tracking
- Download safetensors from HF and convert to unquanted GGUF
Specify multiple KV overrides(added in v1.1.0)Better error handling(added in v1.1.0)Cannot select output/token embd type(fixed in v1.1.0)
Troubleshooting:
- llama.cpp quantizations errors out with an iostream error: create the
quantized_models
directory (or set a directory)