katiyardinesh/DKatiyar-fixed
TEXT GENERATIONConcurrency Cost:1Model Size:8BQuant:FP8Ctx Length:32kPublished:Mar 29, 2026Architecture:Transformer Cold
DKatiyar-fixed is an 8 billion parameter causal language model, based on the Qwen3 architecture, with a 32768 token context length. Developed by katiyardinesh, this model is a corrected version of a previous release, specifically addressing a missing chat template and an incorrect base model attribution. It is primarily designed for chat-based applications, leveraging the ChatML format for structured conversations and tool calls.
Loading preview...
DKatiyar-fixed: A Corrected Qwen3-8B Model
DKatiyar-fixed is an 8 billion parameter language model, developed by katiyardinesh, that resolves critical issues present in a prior release. This model is based on the Qwen3 architecture, not Meta-Llama-3.1 as previously misidentified, and supports a context length of 32768 tokens.
Key Capabilities & Fixes
- Corrected Chat Template: The primary fix involves adding the essential
chat_templatetotokenizer_config.json. This enables proper formatting of role-based messages (system, user, assistant) using the expected<|im_start|>and<|im_end|>tokens, crucial for coherent chat interactions. The template also supports advanced features like tool calls (<tool_call>/</tool_call>) and a thinking mode (<think>/</think>). - Accurate Base Model Identification: The
config.jsonconfirms the model's architecture asQwen3ForCausalLM, aligning with the official Qwen/Qwen3-8B specifications, including its 36 layers, 4096 hidden size, and 151936 vocabulary size.
Good For
- Chat Applications: Ideal for conversational AI systems that require structured, role-based message formatting.
- Tool-Use Scenarios: Supports explicit tool call and thinking modes, making it suitable for agents or applications requiring external function execution.
- Developers Needing a Stable Qwen3-8B Variant: Provides a reliable and correctly configured version of a Qwen3-8B model for deployment.