--- quantized_by: k4yt3x base_model: - k4yt3x/Arynia-LLaMA-70B library_name: transformers tags: - mergekit - merge license: llama3.3 --- This repo contains the GGUF imatrix quants for [k4yt3x/Arynia-LLaMA-70B](https://huggingface.co/k4yt3x/Arynia-LLaMA-70B). The imatrx was computed from [bartowski](https://huggingface.co/bartowski)'s [calibration_datav3.txt](https://gist.github.com/bartowski1182/eb213dccb3571f863da82e99418f81e8) The following precisions are available: - [IQ4_XS](https://huggingface.co/k4yt3x/Arynia-LLaMA-70B-GGUF/resolve/main/Arynia-LLaMA-70B-IQ4_XS.gguf) (37.9 GB) - [Q4_K_M](https://huggingface.co/k4yt3x/Arynia-LLaMA-70B-GGUF/resolve/main/Arynia-LLaMA-70B-Q4_K_M.gguf) (42.5 GB) - [Q5_K_M](https://huggingface.co/k4yt3x/Arynia-LLaMA-70B-GGUF/resolve/main/Arynia-LLaMA-70B-Q5_K_M.gguf) (49.9 GB)