IQ3_XS quant not visible in sidebar
Oh, ok!
I've honestly noticed IQ3_XS/IQ3_XXS missing from a lot of these sidebars in your repos, e.g. https://huggingface.co/bartowski/Qwen2.5-32B-Instruct-GGUF, https://huggingface.co/bartowski/Meta-Llama-3.1-70B-Instruct-GGUF, https://huggingface.co/bartowski/gemma-2-27b-it-GGUF (IQ3_XXS is visible, but IQ3_XS is not) (just 3 random ones)
It's weird. Thank you for looking into this
oh that's even stranger, i assumed they didn't show up anywhere !
Oh wait, I think IQ3_XS is missing everywhere, but IQ3_XXS shows up
I think IQ3_XS doesn't show up for other people either: https://huggingface.co/mradermacher/ThinkPhi1.1-Tensors-i1-GGUF, https://huggingface.co/mradermacher/Qwenvergence-14B-v11-i1-GGUF
Seems like HF has a strange aversion to IQ3_XS
Can you make an IQ3_XXS version of this so it shows up and 16GB VRAM people can use it? Thanks!
I personally do not understand the expediency of such quantization as IQ3_XXS. A model with fewer parameters, but in adequate quantization (>=Q4_K_M) will produce much better answers.