didula-wso2/Qwen3-8B_julia_planning-ep2sft_16bit_vllm
TEXT GENERATIONConcurrency Cost:1Model Size:8BQuant:FP8Ctx Length:32kPublished:Mar 24, 2026License:apache-2.0Architecture:Transformer Open Weights Cold

The didula-wso2/Qwen3-8B_julia_planning-ep2sft_16bit_vllm is an 8 billion parameter Qwen3 model developed by didula-wso2, fine-tuned from didula-wso2/Qwen3-8B_julia_alpaca_ep4sft_16bit_vllm. This model was trained using Unsloth and Huggingface's TRL library, achieving 2x faster training. With a 32768 token context length, it is optimized for specific planning tasks, building upon its base Julia Alpaca fine-tuning.

Loading preview...