yunjae-won/mpq3_qwen4bi_sft_dpo_beta1e-1_step6144

TEXT GENERATIONConcurrency Cost:1Model Size:4BQuant:BF16Ctx Length:32kPublished:Apr 6, 2026Architecture:Transformer Cold

The yunjae-won/mpq3_qwen4bi_sft_dpo_beta1e-1_step6144 is a 4 billion parameter language model with a 32768 token context length. This model is a fine-tuned version, likely based on the Qwen architecture, and is designed for general language generation tasks. Its specific differentiators and primary use cases are not detailed in the provided information.

Loading preview...

Model Overview

The yunjae-won/mpq3_qwen4bi_sft_dpo_beta1e-1_step6144 is a 4 billion parameter language model, featuring a substantial context length of 32768 tokens. This model has undergone supervised fine-tuning (SFT) and DPO (Direct Preference Optimization), indicating an intent to align its outputs with human preferences and instructions.

Key Characteristics

  • Parameter Count: 4 billion parameters, offering a balance between performance and computational efficiency.
  • Context Length: Supports a large context window of 32768 tokens, enabling the model to process and generate longer, more coherent texts.
  • Training Methodology: Utilizes supervised fine-tuning (SFT) and Direct Preference Optimization (DPO), suggesting an emphasis on instruction following and generating preferred responses.

Limitations and Further Information

The provided model card indicates that specific details regarding the model's architecture, training data, intended uses, biases, risks, and evaluation results are currently marked as "More Information Needed." Users should exercise caution and conduct their own assessments regarding its suitability for specific applications until further documentation is made available. The model's direct and downstream uses, as well as its performance benchmarks, are not yet specified.