b44ken/sotu4b

Hugging Face
TEXT GENERATIONConcurrency Cost:1Model Size:4BQuant:BF16Ctx Length:32kPublished:Feb 26, 2026License:apache-2.0Architecture:Transformer Open Weights Warm

The b44ken/sotu4b is a 4 billion parameter Qwen3-based causal language model developed by b44ken. This model was fine-tuned using Unsloth and Hugging Face's TRL library, enabling 2x faster training. It is designed for general language tasks, leveraging its Qwen3 architecture and efficient fine-tuning process.

Loading preview...

Model Overview

The b44ken/sotu4b is a 4 billion parameter Qwen3-based causal language model, fine-tuned by b44ken. This model leverages the Qwen3 architecture and was specifically trained using Unsloth and Hugging Face's TRL library. This combination facilitated a 2x faster fine-tuning process compared to standard methods.

Key Characteristics

  • Architecture: Based on the Qwen3 model family.
  • Parameter Count: 4 billion parameters, offering a balance between performance and computational efficiency.
  • Training Efficiency: Fine-tuned with Unsloth, resulting in significantly faster training times.
  • Context Length: Supports a context window of 32768 tokens.

Potential Use Cases

Given its Qwen3 foundation and efficient fine-tuning, b44ken/sotu4b is suitable for a variety of general language generation and understanding tasks. Its optimized training process suggests it could be a good candidate for applications requiring rapid deployment or iteration on fine-tuned models.