seoyeong903/react_deepseek_1.5B
TEXT GENERATIONConcurrency Cost:1Model Size:1.5BQuant:BF16Ctx Length:32kPublished:Apr 28, 2025Architecture:Transformer Cold

The seoyeong903/react_deepseek_1.5B model is a 1.5 billion parameter language model with a 32768 token context length. This model is a Hugging Face Transformers model, automatically pushed to the Hub. Further details regarding its specific architecture, training data, and primary differentiators are not provided in the available model card.

Loading preview...