moogician/DSR1-Qwen-32B-131fad2c
TEXT GENERATIONConcurrency Cost:2Model Size:32BQuant:FP8Ctx Length:32kLicense:otherArchitecture:Transformer Cold

moogician/DSR1-Qwen-32B-131fad2c is a 32 billion parameter language model, fine-tuned from deepseek-ai/DeepSeek-R1-Distill-Qwen-32B. This model was specifically fine-tuned on the fc_rlm dataset, suggesting an optimization for tasks related to reinforcement learning from human feedback or similar data. It maintains a substantial context length of 32768 tokens, making it suitable for processing extensive inputs.

Loading preview...