joaomdaltoe/me-qwen2.5-1.5B-sft

Hugging Face
TEXT GENERATIONConcurrency Cost:1Model Size:1.5BQuant:BF16Ctx Length:32kPublished:Jan 19, 2026License:apache-2.0Architecture:Transformer Open Weights Warm

The joaomdaltoe/me-qwen2.5-1.5B-sft model is a 1.5 billion parameter causal language model, derived from a merge of a base model and a LoRA/QLoRA adaptation of Qwen/Qwen2.5-1.5B-Instruct. This model maintains the Qwen2.5 architecture and is instruction-tuned, making it suitable for general-purpose conversational AI and text generation tasks. Its compact size allows for efficient deployment while leveraging the capabilities of the Qwen2.5 family.

Loading preview...

Overview

The joaomdaltoe/me-qwen2.5-1.5B-sft model is a 1.5 billion parameter instruction-tuned language model. It is the result of a merge operation (base model + LoRA/QLoRA) applied to the Qwen/Qwen2.5-1.5B-Instruct architecture. This process aims to enhance or adapt the original Qwen2.5-1.5B-Instruct model for specific applications or performance characteristics, while retaining its core capabilities.

Key Capabilities

  • Instruction Following: Designed to respond effectively to user instructions, making it suitable for conversational agents and task-oriented applications.
  • Text Generation: Capable of generating coherent and contextually relevant text across various prompts.
  • Compact Size: With 1.5 billion parameters, it offers a balance between performance and computational efficiency, making it viable for environments with limited resources.

Good for

  • Lightweight Deployment: Ideal for applications requiring a smaller, faster model without sacrificing essential instruction-following abilities.
  • General-Purpose AI: Suitable for a wide range of text-based tasks, including chatbots, content creation, and summarization.
  • Experimentation: Provides a solid base for further fine-tuning or research into LoRA/QLoRA merged models based on the Qwen2.5 family.