djedDJED/qwen7b-lora-r16-lr2e-4-ep4-bf16
TEXT GENERATIONConcurrency Cost:1Model Size:7.6BQuant:FP8Ctx Length:32kPublished:Mar 28, 2026Architecture:Transformer Cold

The djedDJED/qwen7b-lora-r16-lr2e-4-ep4-bf16 model is a 7.6 billion parameter language model, fine-tuned from the Qwen family of models. This model utilizes LoRA (Low-Rank Adaptation) with a rank of 16, trained with a learning rate of 2e-4 over 4 epochs using BF16 precision. It is designed for general language understanding and generation tasks, leveraging its Qwen base for robust performance across various applications.

Loading preview...