LorenaYannnnn/general_reward-Qwen3-0.6B-OURS_llama-seed_2
TEXT GENERATIONConcurrency Cost:1Model Size:0.8BQuant:BF16Ctx Length:32kPublished:Mar 18, 2026Architecture:Transformer Warm

The LorenaYannnnn/general_reward-Qwen3-0.6B-OURS_llama-seed_2 is a 0.8 billion parameter language model based on the Qwen3 architecture. This model is a reward model, likely fine-tuned for evaluating and scoring responses from other language models. Its primary use case is to provide feedback for reinforcement learning from human feedback (RLHF) processes, guiding the training of generative AI models.

Loading preview...