DQN-Labs/dqnagent_v0.1_16bit
TEXT GENERATIONConcurrency Cost:1Model Size:3.8BQuant:BF16Ctx Length:32kPublished:Mar 19, 2026License:apache-2.0Architecture:Transformer Open Weights Cold

DQN-Labs/dqnagent_v0.1_16bit is a 3.8 billion parameter instruction-tuned model developed by DQN-Labs, finetuned from unsloth/phi-4-mini-instruct-unsloth-bnb-4bit. This model is designed for general instruction following tasks, leveraging its 32768 token context length for processing longer inputs. It provides a compact yet capable solution for various natural language processing applications.

Loading preview...

Model Overview

DQN-Labs/dqnagent_v0.1_16bit is a 3.8 billion parameter language model developed by DQN-Labs. It has been instruction-tuned, building upon the unsloth/phi-4-mini-instruct-unsloth-bnb-4bit base model, and is licensed under Apache-2.0. This model is configured with a substantial context length of 32768 tokens, enabling it to handle extensive conversational histories or detailed documents.

Key Capabilities

  • Instruction Following: Designed to accurately interpret and execute a wide range of user instructions.
  • Extended Context Handling: Benefits from a 32768 token context window, suitable for tasks requiring long-form understanding and generation.
  • Efficient Deployment: As a 3.8 billion parameter model, it offers a balance between performance and computational efficiency.

Good For

  • General NLP Tasks: Suitable for common applications like summarization, question answering, and text generation.
  • Applications Requiring Longer Inputs: Ideal for scenarios where the model needs to process and respond to detailed prompts or extended conversations.
  • Resource-Conscious Environments: Its parameter count makes it a viable option for deployment where larger models might be too demanding.