daman1209arora/alpha_0.1_DeepSeek-R1-Distill-Qwen-7B
TEXT GENERATIONConcurrency Cost:1Model Size:7.6BQuant:FP8Ctx Length:32kPublished:Apr 13, 2025Architecture:Transformer Cold

The daman1209arora/alpha_0.1_DeepSeek-R1-Distill-Qwen-7B is a 7.6 billion parameter language model, likely a distilled version based on the DeepSeek-R1 and Qwen architectures. With a substantial context length of 131072 tokens, it is designed to process and generate extensive text sequences. While specific differentiators are not detailed, its architecture suggests a focus on efficient performance for tasks requiring deep contextual understanding.

Loading preview...