From 18d33612e41fec3201dae58212c7e8aaa34a37c1 Mon Sep 17 00:00:00 2001 From: leafspark <78000825+leafspark@users.noreply.github.com> Date: Sun, 4 Aug 2024 16:31:54 -0700 Subject: [PATCH] Created FAQ (markdown) --- FAQ.md | 11 +++++++++++ 1 file changed, 11 insertions(+) create mode 100644 FAQ.md diff --git a/FAQ.md b/FAQ.md new file mode 100644 index 0000000..5867b6a --- /dev/null +++ b/FAQ.md @@ -0,0 +1,11 @@ +1. Q: What is the difference between quantization types? + + A: Different quantization types offer various tradeoffs between model size and inference quality. IQ1_S is the smallest but has worst quality and quantization time, while Q8_0 offers better quality but larger file size and faster quantization. + +2. Q: Can I quantize any HuggingFace model? + + A: Most HuggingFace models compatible with the GGUF format can be quantized using AutoGGUF, although you need to first convert it with the command `python convert_hf_to_gguf.py --outtype auto path_to_your_hf_model` and then move the GGUF to the `models` folder. + +3. Q: How long does quantization take? + + A: Quantization time depends on the model size, quantization type, and your hardware. It can range from minutes to several hours. \ No newline at end of file