ljcamargo/Akkadian-Pretrain-Qwen3-4B-Merged-16B

Hugging Face
TEXT GENERATIONConcurrency Cost:1Model Size:4BQuant:BF16Ctx Length:32kPublished:Mar 20, 2026License:apache-2.0Architecture:Transformer Open Weights Warm

The ljcamargo/Akkadian-Pretrain-Qwen3-4B-Merged-16B is a 4 billion parameter Qwen3 causal language model developed by ljcamargo. This model was finetuned using Unsloth and Huggingface's TRL library, enabling 2x faster training. It is designed for general language tasks, leveraging its Qwen3 architecture and efficient training methodology.

Loading preview...

Model Overview

The ljcamargo/Akkadian-Pretrain-Qwen3-4B-Merged-16B is a 4 billion parameter Qwen3 model developed by ljcamargo. It was finetuned from the unsloth/qwen3-4b-unsloth-bnb-4bit base model, utilizing the Unsloth library in conjunction with Huggingface's TRL library. This combination allowed for significantly faster training, specifically achieving a 2x speed improvement.

Key Characteristics

  • Architecture: Qwen3
  • Parameters: 4 billion
  • Training Efficiency: Leverages Unsloth for 2x faster finetuning.
  • License: Apache-2.0

Intended Use Cases

This model is suitable for a variety of general-purpose language generation and understanding tasks, benefiting from its efficient training and the capabilities of the Qwen3 architecture. Its optimized training process makes it a practical choice for applications requiring a moderately sized, performant language model.