self-hosted api
- run it with
gguf-connector; activate the backend in console/terminal by
ggc w8
- choose your model* file
*accept sd3.5 2b model gguf recently, this will give you the fastest experience for even low tier gpu; frontend https://test.gguf.org or localhost (see decentralized frontend section below)GGUF available. Select which one to use:
Enter your choice (1 to 2): _
- or opt fastapi lumina connector
ggc w7
choose your model* file
GGUF available. Select which one to use:
Enter your choice (1 to 2): _
*as lumina is no lite version recently, might need to increase the step to around 25 for better output
or opt fastapi flux connector
ggc w6
- choose your model* file
*accept any flux model gguf, lite is recommended for saving loading timeGGUF available. Select which one to use:
Enter your choice (1 to 2): _
- flexible frontend choice (see below)
decentralized frontend
- option 1: navigate to https://test.gguf.org
- option 2: localhost; keep the backend running and open a new terminal session then execute
ggc b

- Prompt
- a cat in a hat

- Prompt
- a raccoon in a hat

- Prompt
- a raccoon in a hat

- Prompt
- a dog walking in a cyber city with joy

- Prompt
- a dog walking in a cyber city with joy

- Prompt
- a dog walking in a cyber city with joy
self-hosted api (edit)
- run it with
gguf-connector; activate the backend in console/terminal by
ggc e8
- choose your model file
GGUF available. Select which one to use:
- flux-kontext-lite-q2_k.gguf [4.08GB]
Enter your choice (1 to 1): _
decentralized frontend - opt Edit from pulldown menu (stage 1: exclusive for π· holder trial recently)
- option 1: navigate to https://gguf.org
- option 2: localhost; keep the backend running and open a new terminal session then execute
ggc a
self-hosted api (plus)
- run it with
gguf-connector; activate the backend in console/terminal by
ggc e9
- choose your model file
Safetensors available. Select which one to use:
- sketch-s9-20b-fp4.safetensors (for blackwell card 11.9GB)
- sketch-s9-20b-int4.safetensors (for non-blackwell card 11.5GB)
Enter your choice (1 to 2): _
decentralized frontend - opt Plus from pulldown menu (stage 1: exclusive for π· holder trial recently)
- option 1: navigate to https://gguf.org
- option 2: localhost; keep the backend running and open a new terminal session then execute
ggc a
Inference Providers
NEW
This model isn't deployed by any Inference Provider.
π
Ask for provider support









