rl-rag/qwen3-8B-sft-mix-v20250921
TEXT GENERATIONConcurrency Cost:1Model Size:8BQuant:FP8Ctx Length:32kPublished:Sep 21, 2025License:otherArchitecture:Transformer Cold
The rl-rag/qwen3-8B-sft-mix-v20250921 is an 8 billion parameter language model, fine-tuned from the Qwen/Qwen3-8B architecture. This model has been specifically trained on the rl-rag/sft-mix-v20250921 dataset, indicating an optimization for specific instruction-following or mixed-task performance. With a context length of 32768 tokens, it is designed for applications requiring processing of moderately long inputs and generating coherent, relevant outputs based on its specialized fine-tuning.
Loading preview...