Prashasst/Luffy-DeepSeek-R1-Distill-Llama-8B-4-bit
TEXT GENERATIONConcurrency Cost:1Model Size:8BQuant:FP8Ctx Length:32kPublished:Feb 3, 2025License:mitArchitecture:Transformer0.0K Open Weights Warm

Prashasst/Luffy-DeepSeek-R1-Distill-Llama-8B-4-bit is an 8 billion parameter language model developed by Prashasst Dongre. This model is a distilled version of DeepSeek-R1-Distill-Llama, leveraging a 4-bit quantization for efficient deployment. It is based on the Llama architecture and is designed for general language understanding and generation tasks. With a context length of 32768 tokens, it can process extensive inputs for various applications.

Loading preview...