Prashasst/Luffy-DeepSeek-R1-Distill-Llama-8B-4-bit is an 8 billion parameter language model developed by Prashasst Dongre. This model is a distilled version of DeepSeek-R1-Distill-Llama, leveraging a 4-bit quantization for efficient deployment. It is based on the Llama architecture and is designed for general language understanding and generation tasks. With a context length of 32768 tokens, it can process extensive inputs for various applications.
No reviews yet. Be the first to review!