koutch/qwenb_qwen3-8b_train_sft_train_code
TEXT GENERATIONConcurrency Cost:1Model Size:8BQuant:FP8Ctx Length:32kPublished:Feb 5, 2026License:apache-2.0Architecture:Transformer Open Weights Cold
The koutch/qwenb_qwen3-8b_train_sft_train_code model is an 8 billion parameter Qwen3-based causal language model, fine-tuned by koutch. This model was trained using Unsloth and Huggingface's TRL library, enabling faster training. It is designed for general language tasks, leveraging its Qwen3 architecture and 32768 token context length.
Loading preview...
Model Overview
The koutch/qwenb_qwen3-8b_train_sft_train_code is an 8 billion parameter language model developed by koutch. It is based on the Qwen3 architecture and was fine-tuned from unsloth/qwen3-8b-unsloth-bnb-4bit.
Key Characteristics
- Architecture: Qwen3-based causal language model.
- Parameter Count: 8 billion parameters.
- Training Efficiency: Fine-tuned with Unsloth and Huggingface's TRL library, which facilitated a 2x faster training process.
- Context Length: Supports a context length of 32768 tokens.
Intended Use
This model is suitable for a variety of general language generation and understanding tasks, benefiting from its Qwen3 foundation and efficient fine-tuning methodology.