manhcuong2005/qwen2.5-1.5b-legal-edu-v3

TEXT GENERATIONConcurrency Cost:1Model Size:1.5BQuant:BF16Ctx Length:32kPublished:Apr 17, 2026License:apache-2.0Architecture:Transformer Open Weights Cold

The manhcuong2005/qwen2.5-1.5b-legal-edu-v3 is a 1.5 billion parameter Qwen2.5-based instruction-tuned causal language model developed by manhcuong2005, fine-tuned from unsloth/qwen2.5-1.5b-instruct-unsloth-bnb-4bit. It features a 32768 token context length and was trained using Unsloth and Huggingface's TRL library, enabling 2x faster training. This model is specialized for legal and educational applications, leveraging its efficient training for domain-specific tasks.

Loading preview...

Overview

manhcuong2005/qwen2.5-1.5b-legal-edu-v3 is a 1.5 billion parameter instruction-tuned language model developed by manhcuong2005. It is fine-tuned from the unsloth/qwen2.5-1.5b-instruct-unsloth-bnb-4bit base model and utilizes a substantial 32768 token context length. A key characteristic of this model is its training methodology, which leveraged Unsloth and Huggingface's TRL library, resulting in a 2x acceleration in the training process.

Key Capabilities

  • Domain-Specific Fine-tuning: Optimized for tasks within the legal and educational sectors.
  • Efficient Training: Benefits from Unsloth's accelerated training, making it a potentially cost-effective and faster-to-deploy solution for its target domains.
  • Large Context Window: Supports a 32768 token context, allowing for processing and understanding of extensive documents and conversations relevant to legal and educational content.

Good for

  • Legal Document Analysis: Ideal for tasks requiring understanding and generation of legal texts, such as contract review, case summarization, or legal research assistance.
  • Educational Content Generation: Suitable for creating educational materials, answering academic queries, or assisting in learning platforms.
  • Resource-Efficient Deployment: Its 1.5 billion parameter size, combined with efficient training, makes it a strong candidate for applications where computational resources are a consideration, while still offering specialized domain knowledge.