katiyardinesh/DKatiyar-fixed
TEXT GENERATIONConcurrency Cost:1Model Size:8BQuant:FP8Ctx Length:32kPublished:Mar 29, 2026Architecture:Transformer Cold

DKatiyar-fixed is an 8 billion parameter causal language model, based on the Qwen3 architecture, with a 32768 token context length. Developed by katiyardinesh, this model is a corrected version of a previous release, specifically addressing a missing chat template and an incorrect base model attribution. It is primarily designed for chat-based applications, leveraging the ChatML format for structured conversations and tool calls.

Loading preview...

DKatiyar-fixed: A Corrected Qwen3-8B Model

DKatiyar-fixed is an 8 billion parameter language model, developed by katiyardinesh, that resolves critical issues present in a prior release. This model is based on the Qwen3 architecture, not Meta-Llama-3.1 as previously misidentified, and supports a context length of 32768 tokens.

Key Capabilities & Fixes

  • Corrected Chat Template: The primary fix involves adding the essential chat_template to tokenizer_config.json. This enables proper formatting of role-based messages (system, user, assistant) using the expected <|im_start|> and <|im_end|> tokens, crucial for coherent chat interactions. The template also supports advanced features like tool calls (<tool_call>/</tool_call>) and a thinking mode (<think>/</think>).
  • Accurate Base Model Identification: The config.json confirms the model's architecture as Qwen3ForCausalLM, aligning with the official Qwen/Qwen3-8B specifications, including its 36 layers, 4096 hidden size, and 151936 vocabulary size.

Good For

  • Chat Applications: Ideal for conversational AI systems that require structured, role-based message formatting.
  • Tool-Use Scenarios: Supports explicit tool call and thinking modes, making it suitable for agents or applications requiring external function execution.
  • Developers Needing a Stable Qwen3-8B Variant: Provides a reliable and correctly configured version of a Qwen3-8B model for deployment.