|
--- |
|
license: apache-2.0 |
|
base_model: |
|
- openai/gpt-oss-20b |
|
tags: |
|
- mlx |
|
- gpt |
|
- openai |
|
- chatGPT |
|
--- |
|
|
|
# NexaAI/gpt-oss-20b-MLX-4bit |
|
|
|
## Quickstart |
|
|
|
Run them directly with [nexa-sdk](https://github.com/NexaAI/nexa-sdk) installed |
|
In nexa-sdk CLI: |
|
|
|
```bash |
|
NexaAI/gpt-oss-20b-MLX-4bit |
|
``` |
|
|
|
## Overview |
|
This is a 4-bit quantized version of the OpenAI GPT OSS 20B model, optimized for Apple Silicon using the MLX framework. The model was successfully converted from the original `gpt_oss` architecture to MLX format using the development version of `mlx-lm`. |
|
|
|
|
|
## Reference |
|
**Original model card**: [InferenceIllusionist/gpt-oss-20b-MLX-4bit](https://huggingface.co/InferenceIllusionist/gpt-oss-20b-MLX-4bit) |