Gluwxy/unsloth_finetune
VISIONConcurrency Cost:1Model Size:4.3BQuant:BF16Ctx Length:32kPublished:Jan 7, 2026Architecture:Transformer Cold
Gluwxy/unsloth_finetune is a 4.3 billion parameter Gemma-3.4B-PT model, fine-tuned and converted to GGUF format using Unsloth. This model is optimized for efficient deployment and inference on local hardware, offering various quantization levels. It is specifically designed for text-only LLM applications, with adjusted BOS token behavior for GGUF compatibility.
Loading preview...
Overview
This model, Gluwxy/unsloth_finetune, is a 4.3 billion parameter variant of the Gemma-3.4B-PT architecture. It has been fine-tuned and subsequently converted into the GGUF format using the Unsloth framework, which is noted for enabling faster training.
Key Capabilities
- Efficient Local Deployment: Provided in GGUF format, making it suitable for local inference with tools like
llama.cpp. - Quantization Options: Available in multiple quantization levels (e.g.,
Q5_K_M,Q8_0,Q4_K_M) to balance performance and resource usage. - Text-Only LLM: Primarily designed for text-based language model tasks.
- GGUF Compatibility: The model's Beginning-of-Sentence (BOS) token behavior has been specifically adjusted to ensure proper functionality within the GGUF ecosystem.
Good For
- Developers looking for a Gemma-based model optimized for local execution.
- Applications requiring efficient, quantized models for text generation and understanding.
- Experimentation with models fine-tuned using the Unsloth framework.