Jan-v2-VL-high 4-bit MLX

This is a 4-bit quantized MLX conversion of janhq/Jan-v2-VL-high.

Model Description

Jan-v2-VL is an 8-billion parameter vision-language model designed for long-horizon, multi-step tasks in real software environments. This "high" variant is optimized for deeper reasoning and complex task execution, providing the highest quality outputs for agentic automation and UI control tasks.

Key Features:

  • Vision-language understanding for browser and desktop applications
  • Screenshot grounding and tool call capabilities
  • Stable multi-step execution with minimal performance drift
  • Error recovery and intermediate state maintenance

Quantization

This model was converted to MLX format with 4-bit quantization using MLX-VLM by Prince Canuma.

Conversion command:

mlx_vlm.convert --hf-path janhq/Jan-v2-VL-high --quantize --q-bits 4 --mlx-path Jan-v2-VL-high-4bit-mlx

Usage

Installation

pip install mlx-vlm

Python

from mlx_vlm import load, generate
from mlx_vlm.prompt_utils import apply_chat_template
from mlx_vlm.utils import load_config

# Load the model
model_path = "mlx-community/Jan-v2-VL-high-4bit-mlx"
model, processor = load(model_path)
config = load_config(model_path)

# Prepare input
image = ["path/to/image.jpg"]
prompt = "Describe this image."

# Apply chat template
formatted_prompt = apply_chat_template(
    processor, config, prompt, num_images=len(image)
)

# Generate output
output = generate(model, processor, formatted_prompt, image, verbose=False)
print(output)

Command Line

mlx_vlm.generate --model mlx-community/Jan-v2-VL-high-4bit-mlx --max-tokens 100 --prompt "Describe this image" --image path/to/image.jpg

Intended Use

This model is designed for:

  • Agentic automation and UI control
  • Stepwise operation in browsers and desktop applications
  • Screenshot grounding and tool calls
  • Long-horizon multi-step task execution

License

This model is released under the Apache 2.0 license.

Original Model

For more information, please refer to the original model: janhq/Jan-v2-VL-high

Acknowledgments

Downloads last month
54
Inference Providers NEW
This model isn't deployed by any Inference Provider. ๐Ÿ™‹ Ask for provider support

Model tree for mlx-community/Jan-v2-VL-high-4bit-mlx

Quantized
(8)
this model