smsk1999/qwen3-8b-profiling-merged-v4

TEXT GENERATIONConcurrency Cost:1Model Size:8BQuant:FP8Ctx Length:32kPublished:Apr 28, 2026License:apache-2.0Architecture:Transformer Open Weights Cold

The smsk1999/qwen3-8b-profiling-merged-v4 is an 8 billion parameter Qwen3 model developed by smsk1999, fine-tuned from unsloth/Qwen3-8B-unsloth-bnb-4bit. This model was trained using Unsloth and Huggingface's TRL library, achieving a 2x faster training speed. It is designed for general language tasks, leveraging its efficient training methodology to provide a capable foundation model.

Loading preview...

Model Overview

The smsk1999/qwen3-8b-profiling-merged-v4 is an 8 billion parameter Qwen3 language model developed by smsk1999. It was fine-tuned from the unsloth/Qwen3-8B-unsloth-bnb-4bit base model, utilizing Unsloth and Huggingface's TRL library for training.

Key Characteristics

  • Base Model: Qwen3-8B architecture.
  • Parameter Count: 8 billion parameters.
  • Training Efficiency: Achieved 2x faster training speed through the use of Unsloth.
  • Context Length: Supports a context length of 32768 tokens.
  • License: Distributed under the Apache-2.0 license.

Use Cases

This model is suitable for a variety of general language understanding and generation tasks, benefiting from its efficient fine-tuning process. Its 8B parameter size and substantial context window make it a versatile option for applications requiring robust language capabilities.