RedHatAI/Qwen2-0.5B-Instruct-quantized.w8a8
Text Generation
•
0.6B
•
Updated
•
25
RedHatAI/Phi-3-mini-128k-instruct-quantized.w4a16
Text Generation
•
0.7B
•
Updated
•
10
•
1
RedHatAI/Qwen2-1.5B-Instruct-quantized.w8a8
Text Generation
•
2B
•
Updated
•
135
RedHatAI/Phi-3-mini-128k-instruct-quantized.w8a8
Text Generation
•
4B
•
Updated
•
7
RedHatAI/Meta-Llama-3-8B-Instruct-quantized.w8a8
Text Generation
•
8B
•
Updated
•
3.62k
•
2
RedHatAI/Llama-2-7b-chat-quantized.w8a8
Text Generation
•
7B
•
Updated
•
234
•
1
RedHatAI/Phi-3-mini-128k-instruct-quantized.w8a16
Text Generation
•
1B
•
Updated
•
13
RedHatAI/Phi-3-mini-128k-instruct-FP8
Text Generation
•
4B
•
Updated
•
13
RedHatAI/Llama-3.2-3B-Instruct-FP8-dynamic
Text Generation
•
4B
•
Updated
•
303
•
3
RedHatAI/Llama-3.2-1B-Instruct-FP8-dynamic
Text Generation
•
1B
•
Updated
•
116k
•
3
RedHatAI/gemma-2-9b-it-quantized.w8a8
Text Generation
•
10B
•
Updated
•
9
•
2
RedHatAI/Phi-3-medium-128k-instruct-quantized.w8a8
Text Generation
•
14B
•
Updated
•
20
•
2
RedHatAI/Phi-3-medium-128k-instruct-quantized.w8a16
Text Generation
•
4B
•
Updated
•
6
•
2
RedHatAI/Phi-3-medium-128k-instruct-FP8
Text Generation
•
14B
•
Updated
•
79
•
5
RedHatAI/Qwen2.5-32B-Instruct-quantized.w8a16
9B
•
Updated
•
5
RedHatAI/Qwen2.5-7B-Instruct-quantized.w8a16
3B
•
Updated
•
68
RedHatAI/Qwen2.5-0.5B-Instruct-quantized.w8a16
0.4B
•
Updated
•
8
RedHatAI/Qwen2.5-72B-Instruct-quantized.w8a8
73B
•
Updated
•
88
RedHatAI/Qwen2.5-32B-Instruct-quantized.w8a8
33B
•
Updated
•
5
RedHatAI/Qwen2.5-32B-quantized.w8a8
33B
•
Updated
•
4
RedHatAI/Meta-Llama-3.1-405B-Instruct-FP8
Text Generation
•
406B
•
Updated
•
1.04k
•
31
RedHatAI/Qwen2.5-3B-Instruct-quantized.w8a8
3B
•
Updated
•
10
RedHatAI/Qwen2.5-1.5B-Instruct-quantized.w8a8
2B
•
Updated
•
5
RedHatAI/SparseLlama-3-8B-pruned_50.2of4
Text Generation
•
8B
•
Updated
•
7
RedHatAI/Llama-3.2-90B-Vision-Instruct-FP8-dynamic
Text Generation
•
89B
•
Updated
•
262k
•
10
RedHatAI/Llama-3.2-11B-Vision-Instruct-FP8-dynamic
Text Generation
•
11B
•
Updated
•
557
•
24
RedHatAI/Phi-3.5-mini-instruct-FP8-KV
Text Generation
•
4B
•
Updated
•
14
•
2
RedHatAI/Meta-Llama-3-70B-Instruct-quantized.w4a16
Text Generation
•
11B
•
Updated
•
74
•
2
RedHatAI/SmolLM-135M-q
Updated
RedHatAI/Mixtral-8x22B-Instruct-v0.1-AutoFP8
Text Generation
•
141B
•
Updated
•
8
•
3