sming256/Qwen3-VL-8B-Instruct-Patched

VISIONConcurrency Cost:1Model Size:8BQuant:FP8Ctx Length:32kPublished:Dec 31, 2025License:apache-2.0Architecture:Transformer Open Weights Cold

The sming256/Qwen3-VL-8B-Instruct-Patched is an 8 billion parameter vision-language model from the Qwen series, developed by Qwen. This patched variant of Qwen3-VL-8B-Instruct features removed and tokens to prevent tokenizer failures during GRPO. It offers enhanced visual perception, reasoning, and agent interaction capabilities, excelling in multimodal tasks with a 32768 token context length.

Loading preview...

Qwen3-VL-8B-Instruct-Patched: An Advanced Vision-Language Model

This model is a patched version of the Qwen3-VL-8B-Instruct, specifically modified to remove <think> and </think> tokens from its tokenizer configuration. This ensures robust performance during GRPO by preventing tokenizer failures when these tokens appear in prompts or outputs.

Key Capabilities and Enhancements:

  • Comprehensive Multimodal Understanding: Offers superior text understanding and generation, combined with deep visual perception and reasoning.
  • Visual Agent Functionality: Capable of operating PC/mobile GUIs, recognizing elements, understanding functions, and completing tasks.
  • Visual Coding Boost: Generates Draw.io, HTML, CSS, and JS from image and video inputs.
  • Advanced Spatial Perception: Judges object positions, viewpoints, and occlusions, enabling stronger 2D and 3D spatial reasoning.
  • Long Context & Video Understanding: Features a native 256K context, expandable to 1M, allowing it to process extensive documents and hours-long video content with detailed recall.
  • Enhanced Multimodal Reasoning: Excels in STEM and Math tasks, providing causal analysis and logical, evidence-based answers.
  • Upgraded Visual Recognition: Benefits from broader and higher-quality pretraining, enabling recognition of a vast array of entities including celebrities, anime, products, and landmarks.
  • Expanded OCR: Supports 32 languages and demonstrates robustness in challenging conditions like low light, blur, and tilt, with improved handling of rare characters and long document structures.
  • Seamless Text-Vision Fusion: Achieves text understanding on par with pure LLMs through unified, lossless comprehension across modalities.

Architectural Updates:

  • Interleaved-MRoPE: Utilizes robust positional embeddings for full-frequency allocation across time, width, and height, enhancing long-horizon video reasoning.
  • DeepStack: Fuses multi-level ViT features to capture fine-grained details and improve image-text alignment.
  • Text-Timestamp Alignment: Employs precise, timestamp-grounded event localization for stronger video temporal modeling.

This model is ideal for applications requiring advanced multimodal interaction, visual coding, and comprehensive understanding of both static images and dynamic video content.