automatically quant GGUF models
Go to file
BuildTools 868b15b61f ci: add pre commit config and report template 2024-08-04 20:59:02 -07:00
.github ci: add pre commit config and report template 2024-08-04 20:59:02 -07:00
assets edit favicon 2024-08-04 16:04:04 -07:00
src ci: add pre commit config and report template 2024-08-04 20:59:02 -07:00
.gitattributes chore: add .gitattributes 2024-08-04 18:52:14 -07:00
.gitignore ci: add pre commit config and report template 2024-08-04 20:59:02 -07:00
.pre-commit-config.yaml ci: add pre commit config and report template 2024-08-04 20:59:02 -07:00
CHANGELOG.md docs: update changelog to v1.4.2 2024-08-04 19:37:19 -07:00
CONTRIBUTING.md docs: update formatting instructions 2024-08-04 19:47:29 -07:00
LICENSE add details 2024-08-03 19:41:08 -07:00
README.md docs: update readme to v1.4.2 2024-08-04 19:45:01 -07:00
build.bat edit favicon 2024-08-04 16:04:04 -07:00
build.sh build: add cross-platform build scripts 2024-08-04 19:04:02 -07:00
build_nfo.bat build: add cross-platform build scripts 2024-08-04 19:04:02 -07:00
build_nfo.sh build: add cross-platform build scripts 2024-08-04 19:04:02 -07:00
requirements.txt fix: Hide Base Model label when GGML type selected 2024-08-04 19:32:07 -07:00
run.bat modify backend check logic 2024-08-04 09:12:07 -07:00

README.md

AutoGGUF-banner

AutoGGUF - automated GGUF model quantizer

GitHub release GitHub last commit CI/CD Status

Powered by llama.cpp GitHub top language Platform Compatibility GitHub license

GitHub stars GitHub forks GitHub repo size

Code Style: Black Issues PRs Welcome

AutoGGUF provides a graphical user interface for quantizing GGUF models using the llama.cpp library. It allows users to download different versions of llama.cpp, manage multiple backends, and perform quantization tasks with various options.

Features

  • Download and manage llama.cpp backends
  • Select and quantize GGUF models
  • Configure quantization parameters
  • Monitor system resources during quantization

Usage

Cross-platform

  1. Install dependencies:
    pip install -r requirements.txt
    
    or
    pip install PyQt6 requests psutil shutil
    
  2. Run the application:
    python src/main.py
    
    or use the run.bat script.

Windows

  1. Download the latest release
  2. Extract all files to a folder
  3. Run AutoGGUF.exe

Building

Cross-platform

cd src
pip install -U pyinstaller
pyinstaller main.py --onefile
cd dist/main
./main

Windows

build RELEASE | DEV

Find the executable in build/<type>/dist/AutoGGUF.exe.

Dependencies

  • PyQt6
  • psutil
  • shutil
  • numpy
  • torch
  • safetensors
  • gguf (bundled)

Localizations

View the list of supported languages at AutoGGUF/wiki/Installation#configuration (LLM translated, except for English).

To use a specific language, set the AUTOGGUF_LANGUAGE environment variable to one of the listed language codes.

Known Issues

  • Saving preset while quantizing causes UI thread crash (planned fix: remove this feature)
  • Cannot delete task while processing (planned fix: disallow deletion before cancelling or cancel automatically)
  • Base Model text still shows when GGML is selected as LoRA type (fix: include text in show/hide Qt layout) (fixed in v1.4.2)

Planned Features

  • Actual progress bar tracking
  • Download safetensors from HF and convert to unquantized GGUF
  • Perplexity testing
  • Managing shards (coming in the next release)
  • Time estimation for quantization
  • Dynamic values for KV cache (coming in the next release)
  • Ability to select and start multiple quants at once (saved in presets, coming in the next release)

Troubleshooting

  • SSL module cannot be found error: Install OpenSSL or run from source using python src/main.py with the run.bat script (pip install requests)

Contributing

Fork the repo, make your changes, and ensure you have the latest commits when merging. Include a changelog of new features in your pull request description. Read CONTRIBUTING.md for more information.

User Interface

image

Stargazers

Star History Chart