PatrickChikuse/chichewa-agri-qwen

TEXT GENERATIONConcurrency Cost:1Model Size:1.5BQuant:BF16Ctx Length:32kPublished:Apr 29, 2026License:apache-2.0Architecture:Transformer Open Weights Cold

The PatrickChikuse/chichewa-agri-qwen is a 1.5 billion parameter Qwen2.5-based instruction-tuned causal language model developed by PatrickChikuse. Finetuned from unsloth/qwen2.5-1.5b-instruct-unsloth-bnb-4bit, it leverages Unsloth and Huggingface's TRL library for accelerated training. This model is specifically designed for applications requiring a compact yet capable language model, potentially optimized for agricultural contexts in Chichewa, given its name. It offers a 32768 token context length, making it suitable for processing longer inputs.

Loading preview...

Model Overview

The PatrickChikuse/chichewa-agri-qwen is a 1.5 billion parameter instruction-tuned language model, developed by PatrickChikuse. It is built upon the Qwen2.5 architecture and was finetuned from the unsloth/qwen2.5-1.5b-instruct-unsloth-bnb-4bit base model.

Key Characteristics

  • Architecture: Qwen2.5-based, a robust causal language model family.
  • Parameter Count: 1.5 billion parameters, offering a balance between performance and computational efficiency.
  • Context Length: Supports a substantial context window of 32768 tokens, enabling the processing of longer texts and complex queries.
  • Training Efficiency: The model was trained using Unsloth and Huggingface's TRL library, which facilitated a 2x faster finetuning process.
  • Potential Specialization: The model's name, "chichewa-agri-qwen," suggests a potential specialization in agricultural topics and the Chichewa language, making it distinct from general-purpose LLMs.

Use Cases

This model is particularly well-suited for applications where:

  • Resource Efficiency is Key: Its 1.5B parameter size makes it suitable for deployment on devices with limited computational resources.
  • Long Context Understanding is Required: The 32k context window allows for processing and generating responses based on extensive input texts.
  • Domain-Specific Applications: If its implied specialization holds true, it would be ideal for tasks related to agriculture, especially within the Chichewa linguistic context.
  • Fast Finetuning is Desired: The use of Unsloth indicates a focus on efficient model development and iteration.