---
library_name: transformers
base_model:
- Qwen/Qwen3-VL-2B-Instruct
pipeline_tag: text-generation
---
# Qwen3-VLTO-1.7B-Instruct
Qwen3-VL-2B-Instruct but without the vision components (**V**ision **L**anguage **T**ext **O**nly). Functions exactly like a text-only Qwen3 model.
To do this, I simply imported the weights from the VL model into the text model via PyTorch's `load_state_dict`. The model architecture is essentially the exact same.