Q6_K vs Q6_K_L
i couldn't find good information about "Q6_K vs Q6_K_L" (speed and accuracy and ...)
all articles and discussions are about "IQ4 vs Q4 vs Q5 vs Q6" π
The only difference is that Q6_K_L uses Q8_0 for embed/output weights instead of Q6_K, so it could have slightly higher performance
In practice it's hard to find much difference, especially for the larger weights. Q2_K vs Q2_K_L is where the only major difference is seen.
Frankly, Q6 is already so close to Q8 that the + _L is pretty much a waste of VRAM and compute time, it's better spent on context length.
yeah i mostly agree, I don't necessarily recommend them over the regular ones, but some people do, and I'm nothing if not a people pleaser...
The only difference is that Q6_K_L uses Q8_0 for embed/output weights instead of Q6_K, so it could have slightly higher performance
In practice it's hard to find much difference, especially for the larger weights. Q2_K vs Q2_K_L is where the only major difference is seen.
Thank you so much ! ππ