SiwaSathya/model
VISIONConcurrency Cost:1Model Size:4.3BQuant:BF16Ctx Length:32kPublished:Jan 2, 2026Architecture:Transformer Cold
SiwaSathya/model is a 4.3 billion parameter language model, likely based on the Gemma architecture given its file names, fine-tuned and converted to GGUF format using Unsloth. It supports a context length of 32768 tokens and includes variants for both text-only and multimodal (vision) applications. This model is optimized for efficient deployment and usage with tools like llama.cpp and Ollama, offering faster training capabilities due to its Unsloth integration.
Loading preview...
Overview
SiwaSathya/model is a 4.3 billion parameter language model, fine-tuned and converted into the GGUF format using the Unsloth library, which enabled 2x faster training. This model is designed for efficient deployment and use with llama.cpp and Ollama.
Key Capabilities
- GGUF Format: Provided in various quantization levels (Q5_K_M, Q8_0, Q4_K_M) for flexible deployment.
- Multimodal Support: Includes a
F16-mmproj.gguffile, indicating vision capabilities, usable withllama.cpp/llama-mtmd-cli. - Optimized for Deployment: Ready for use with
llama.cppviallama-cliorllama-mtmd-cli. - Ollama Compatibility: Specific instructions are provided for creating a unified bf16 model for Ollama, addressing its current lack of separate
mmprojfile support.
Good for
- Developers seeking a 4.3B parameter model in GGUF format for local inference.
- Applications requiring both text-only and multimodal (vision) capabilities.
- Users prioritizing faster training and efficient model conversion via Unsloth.
- Deployment on platforms like
llama.cppandOllama.