Gluwxy/unsloth_finetune
VISIONConcurrency Cost:1Model Size:4.3BQuant:BF16Ctx Length:32kPublished:Jan 7, 2026Architecture:Transformer Cold

Gluwxy/unsloth_finetune is a 4.3 billion parameter Gemma-3.4B-PT model, fine-tuned and converted to GGUF format using Unsloth. This model is optimized for efficient deployment and inference on local hardware, offering various quantization levels. It is specifically designed for text-only LLM applications, with adjusted BOS token behavior for GGUF compatibility.

Loading preview...

Overview

This model, Gluwxy/unsloth_finetune, is a 4.3 billion parameter variant of the Gemma-3.4B-PT architecture. It has been fine-tuned and subsequently converted into the GGUF format using the Unsloth framework, which is noted for enabling faster training.

Key Capabilities

  • Efficient Local Deployment: Provided in GGUF format, making it suitable for local inference with tools like llama.cpp.
  • Quantization Options: Available in multiple quantization levels (e.g., Q5_K_M, Q8_0, Q4_K_M) to balance performance and resource usage.
  • Text-Only LLM: Primarily designed for text-based language model tasks.
  • GGUF Compatibility: The model's Beginning-of-Sentence (BOS) token behavior has been specifically adjusted to ensure proper functionality within the GGUF ecosystem.

Good For

  • Developers looking for a Gemma-based model optimized for local execution.
  • Applications requiring efficient, quantized models for text generation and understanding.
  • Experimentation with models fine-tuned using the Unsloth framework.