koutch/qwen_2.json_train_dpo_v1_train_code
TEXT GENERATIONConcurrency Cost:1Model Size:4BQuant:BF16Ctx Length:32kPublished:Feb 3, 2026License:apache-2.0Architecture:Transformer Open Weights Warm

The koutch/qwen_2.json_train_dpo_v1_train_code model is a 4 billion parameter Qwen3-based causal language model developed by koutch, fine-tuned for code-related tasks. It was trained using Unsloth and Huggingface's TRL library, enabling faster training. This model is optimized for code generation and understanding, leveraging its 40960 token context length for complex programming challenges.

Loading preview...