-
-
-
-
-
-
Inference Providers
Active filters:
gptq
acl-srw-2024/llama-3-typhoon-v1.5-8b-instruct-unsloth-sft-epoch-3-gptq-4bit
Text Generation
•
Updated
•
6
acl-srw-2024/llama-3-typhoon-v1.5-8b-instruct-unsloth-sft-epoch-3-gptq-8bit
Text Generation
•
Updated
•
6
agahebr/magnum-gptq-32gs
Text Generation
•
Updated
•
9
acl-srw-2024/openthai-7b-unsloth-gptq-2bit
Text Generation
•
Updated
•
8
acl-srw-2024/openthai-7b-unsloth-gptq-3bit
Text Generation
•
Updated
•
6
acl-srw-2024/openthai-7b-unsloth-gptq-4bit
Text Generation
•
Updated
•
7
acl-srw-2024/openthai-7b-unsloth-gptq-8bit
Text Generation
•
Updated
•
6
Udith-Sandaruwan/opt-125m-gptq-4bit
Text Generation
•
Updated
•
63
Intern95/opt-125m-gptq
Text Generation
•
Updated
•
63
Intern95/opt-350m-gptq
Text Generation
•
Updated
•
66
acl-srw-2024/SeaLLM-7B-v2.5-unsloth-sft-epoch-3-gptq-2bit
Text Generation
•
Updated
•
6
acl-srw-2024/SeaLLM-7B-v2.5-unsloth-sft-epoch-3-gptq-3bit
Text Generation
•
Updated
•
6
ModelCloud/gemma-2-9b-gptq-4bit
Text Generation
•
Updated
•
15
Granther/Gemma-2-9B-Instruct-4Bit-GPTQ
Text Generation
•
Updated
•
527
•
3
minyichen/Llama-3-Taiwan-70B-Instruct-GPTQ
Text Generation
•
Updated
•
53
•
2
anhduy0911/LLM-IE-Healthcare
Text Generation
•
Updated
•
7
nm-testing/SparseLlama-2-7b-evolcodealpaca-pruned_50.2of4-quantized.w4a16
Text Generation
•
Updated
•
65
gdsaikrishna/Qwen_1.5_Base_gptq_int4_g128_wikitext2
Text Generation
•
Updated
•
63
XavierSpycy/Meta-Llama-3-8B-Instruct-zh-10k-GPTQ
Text Generation
•
Updated
•
18
BlackSamorez/Meta-Llama-3-70B-Instruct-GPTQ
Text Generation
•
Updated
•
15
lfnothing/opt-125m-gptq
Text Generation
•
Updated
•
64
neuralmagic/Llama-2-7b-chat-quantized.w8a16
Text Generation
•
Updated
•
185
neuralmagic/Meta-Llama-3-8B-Instruct-quantized.w8a16
Text Generation
•
Updated
•
34.4k
•
3
ramitpahwa13/OpenHermes-2.5-Mistral-7B-Pruned50-GPTQ-NO-Marlin
Text Generation
•
Updated
•
76
neuralmagic/Qwen2-0.5B-Instruct-quantized.w8a16
Text Generation
•
Updated
•
158
neuralmagic/Qwen2-1.5B-Instruct-quantized.w8a16
Text Generation
•
Updated
•
184
neuralmagic/Qwen2-7B-Instruct-quantized.w8a16
Text Generation
•
Updated
•
33
neuralmagic/Qwen2-72B-Instruct-quantized.w8a16
Text Generation
•
Updated
•
40
•
1
GalaktischeGurke/Tess-v2.5.2-Qwen2-72B-4Bit-GPTQ
Text Generation
•
Updated
•
7
acl-srw-2024/SeaLLM-7B-v2.5-unsloth-sft-epoch-3-gptq-4bit
Text Generation
•
Updated
•
6