hmdmahdavi/s1-generator-critique-Qwen3-4B-Instruct-2507-20251214_200751
TEXT GENERATIONConcurrency Cost:1Model Size:4BQuant:BF16Ctx Length:32kArchitecture:Transformer Warm

This model is a 4 billion parameter instruction-tuned language model, fine-tuned by hmdmahdavi from the Qwen3-4B-Instruct-2507 base model. With a context length of 40960 tokens, it is optimized for generating critiques and responses to open-ended questions. It leverages SFT training via TRL to enhance its conversational and generative capabilities.

Loading preview...