daman1209arora/alpha_0_DeepSeek-R1-Distill-Qwen-7B
TEXT GENERATIONConcurrency Cost:1Model Size:7.6BQuant:FP8Ctx Length:32kPublished:Apr 13, 2025Architecture:Transformer Cold

daman1209arora/alpha_0_DeepSeek-R1-Distill-Qwen-7B is a 7.6 billion parameter language model. Based on the model name, it appears to be a distilled version of DeepSeek-R1 and Qwen-7B architectures. Due to the lack of specific details in its model card, its primary differentiators and specific use cases are not explicitly defined, suggesting it may be a foundational or experimental model requiring further fine-tuning or evaluation.

Loading preview...