Jan-v2-VL-high BF16 MLX
This is a BF16 (bfloat16) precision MLX conversion of janhq/Jan-v2-VL-high.
Model Description
Jan-v2-VL is an 8-billion parameter vision-language model designed for long-horizon, multi-step tasks in real software environments. This "high" variant is optimized for deeper reasoning and complex task execution, providing the highest quality outputs for agentic automation and UI control tasks.
Key Features:
- Vision-language understanding for browser and desktop applications
- Screenshot grounding and tool call capabilities
- Stable multi-step execution with minimal performance drift
- Error recovery and intermediate state maintenance
Precision
This model was converted to MLX format with bfloat16 precision (no quantization) using MLX-VLM by Prince Canuma. BF16 provides near full-precision quality with reduced memory footprint.
Conversion command:
mlx_vlm.convert --hf-path janhq/Jan-v2-VL-high --dtype bfloat16 --mlx-path Jan-v2-VL-high-bf16-mlx
Usage
Installation
pip install mlx-vlm
Python
from mlx_vlm import load, generate
from mlx_vlm.prompt_utils import apply_chat_template
from mlx_vlm.utils import load_config
# Load the model
model_path = "mlx-community/Jan-v2-VL-high-bf16-mlx"
model, processor = load(model_path)
config = load_config(model_path)
# Prepare input
image = ["path/to/image.jpg"]
prompt = "Describe this image."
# Apply chat template
formatted_prompt = apply_chat_template(
processor, config, prompt, num_images=len(image)
)
# Generate output
output = generate(model, processor, formatted_prompt, image, verbose=False)
print(output)
Command Line
mlx_vlm.generate --model mlx-community/Jan-v2-VL-high-bf16-mlx --max-tokens 100 --prompt "Describe this image" --image path/to/image.jpg
Intended Use
This model is designed for:
- Agentic automation and UI control
- Stepwise operation in browsers and desktop applications
- Screenshot grounding and tool calls
- Long-horizon multi-step task execution
License
This model is released under the Apache 2.0 license.
Original Model
For more information, please refer to the original model: janhq/Jan-v2-VL-high
Acknowledgments
- Original model by Jan
- MLX framework by Apple
- MLX conversion framework by Prince Canuma
- Model conversion by Incept5
- Downloads last month
- 115
Inference Providers
NEW
This model isn't deployed by any Inference Provider.
๐
Ask for provider support