staeiou/bartleby-qwen3-1.7b_dpo
TEXT GENERATIONConcurrency Cost:1Model Size:2BQuant:BF16Ctx Length:32kPublished:Mar 26, 2026Architecture:Transformer Warm

The staeiou/bartleby-qwen3-1.7b_dpo model is a 1.7 billion parameter language model, fine-tuned using Direct Preference Optimization (DPO). This model leverages the Qwen3 architecture and is optimized for generating responses aligned with human preferences. It is suitable for tasks requiring nuanced and preferred text generation, building upon its base model's capabilities.

Loading preview...