RJTPP/scot0500s-deepseek-14b-full

TEXT GENERATIONConcurrency Cost:1Model Size:14.8BQuant:FP8Ctx Length:32kPublished:Apr 21, 2026License:apache-2.0Architecture:Transformer Open Weights Cold

RJTPP/scot0500s-deepseek-14b-full is a 14.8 billion parameter Qwen2-based language model, fine-tuned by RJTPP. This model was developed using Unsloth and Huggingface's TRL library, enabling faster training. It is designed for general language tasks, leveraging its Qwen2 architecture for robust performance.

Loading preview...

Model Overview

RJTPP/scot0500s-deepseek-14b-full is a 14.8 billion parameter language model, fine-tuned by RJTPP. It is based on the Qwen2 architecture, specifically finetuned from unsloth/DeepSeek-R1-Distill-Qwen-14B-unsloth-bnb-4bit.

Key Characteristics

  • Architecture: Qwen2-based, leveraging the DeepSeek-R1-Distill foundation.
  • Training Efficiency: Fine-tuned using Unsloth and Huggingface's TRL library, which facilitated a 2x faster training process.
  • Parameters: Features 14.8 billion parameters, offering a balance between performance and computational requirements.
  • Context Length: Supports a context length of 32768 tokens, suitable for processing longer inputs and generating coherent extended outputs.

Potential Use Cases

This model is suitable for a variety of general language understanding and generation tasks, benefiting from its Qwen2 foundation and efficient fine-tuning. Its substantial parameter count and context length make it a strong candidate for applications requiring detailed comprehension and extensive text generation.