trnqphu/deepseek-r1-4b
TEXT GENERATIONConcurrency Cost:1Model Size:1.5BQuant:BF16Ctx Length:32kPublished:Apr 16, 2025Architecture:Transformer Cold

trnqphu/deepseek-r1-4b is a 1.5 billion parameter language model fine-tuned from deepseek-ai/DeepSeek-R1-Distill-Qwen-1.5B, leveraging a 32768 token context length. This model was trained using Supervised Fine-Tuning (SFT) with the TRL framework. It is designed for general text generation tasks, building upon the DeepSeek-R1 architecture.

Loading preview...