SiwaSathya/model

VISIONConcurrency Cost:1Model Size:4.3BQuant:BF16Ctx Length:32kPublished:Jan 2, 2026Architecture:Transformer Cold

SiwaSathya/model is a 4.3 billion parameter language model, likely based on the Gemma architecture given its file names, fine-tuned and converted to GGUF format using Unsloth. It supports a context length of 32768 tokens and includes variants for both text-only and multimodal (vision) applications. This model is optimized for efficient deployment and usage with tools like llama.cpp and Ollama, offering faster training capabilities due to its Unsloth integration.

Loading preview...

Overview

SiwaSathya/model is a 4.3 billion parameter language model, fine-tuned and converted into the GGUF format using the Unsloth library, which enabled 2x faster training. This model is designed for efficient deployment and use with llama.cpp and Ollama.

Key Capabilities

  • GGUF Format: Provided in various quantization levels (Q5_K_M, Q8_0, Q4_K_M) for flexible deployment.
  • Multimodal Support: Includes a F16-mmproj.gguf file, indicating vision capabilities, usable with llama.cpp/llama-mtmd-cli.
  • Optimized for Deployment: Ready for use with llama.cpp via llama-cli or llama-mtmd-cli.
  • Ollama Compatibility: Specific instructions are provided for creating a unified bf16 model for Ollama, addressing its current lack of separate mmproj file support.

Good for

  • Developers seeking a 4.3B parameter model in GGUF format for local inference.
  • Applications requiring both text-only and multimodal (vision) capabilities.
  • Users prioritizing faster training and efficient model conversion via Unsloth.
  • Deployment on platforms like llama.cpp and Ollama.