roleplaiapp/R1-Qwen2.5-32B-Instruct-1k-i1-IQ3_XS-GGUF

Repo: roleplaiapp/R1-Qwen2.5-32B-Instruct-1k-i1-IQ3_XS-GGUF Original Model: R1-Qwen2.5-32B-Instruct-1k-i1 Quantized File: R1-Qwen2.5-32B-Instruct-1k.i1-IQ3_XS.gguf Quantization: GGUF Quantization Method: IQ3_XS

Overview

This is a GGUF IQ3_XS quantized version of R1-Qwen2.5-32B-Instruct-1k-i1

Quantization By

I often have idle GPUs while building/testing for the RP app, so I put them to use quantizing models. I hope the community finds these quantizations useful.

Andrew Webby @ RolePlai.

Downloads last month
12
GGUF
Model size
32.8B params
Architecture
qwen2
Hardware compatibility
Log In to view the estimation
Inference Providers NEW
This model isn't deployed by any Inference Provider. ๐Ÿ™‹ Ask for provider support