TeichAI/Qwen3-4B-Thinking-2507-Gemini-2.5-Flash-Lite-Preview-Distill
TEXT GENERATIONConcurrency Cost:1Model Size:4BQuant:BF16Ctx Length:32kLicense:apache-2.0Architecture:Transformer0.0K Open Weights Warm
TeichAI/Qwen3-4B-Thinking-2507-Gemini-2.5-Flash-Lite-Preview-Distill is a Qwen3-based language model developed by TeichAI. This model was fine-tuned using 1000 examples from the Gemini 2.5 Flash Lite Preview 09-2025 dataset. It leverages Unsloth and Huggingface's TRL library for accelerated training. This model is designed for general text generation tasks, benefiting from its distillation from a powerful source model.
Loading preview...
Overview
TeichAI/Qwen3-4B-Thinking-2507-Gemini-2.5-Flash-Lite-Preview-Distill is a language model developed by TeichAI, based on the Qwen3 architecture. It was fine-tuned from the unsloth/Qwen3-4B-Thinking-2507 base model.
Key Capabilities
- Distilled Knowledge: The model was trained on 1000 examples sourced from the Gemini 2.5 Flash Lite Preview 09-2025 dataset, aiming to distill capabilities from a more advanced model.
- Efficient Training: Training was conducted using Unsloth and Huggingface's TRL library, enabling a 2x faster fine-tuning process.
- Apache 2.0 License: The model is released under the permissive Apache 2.0 license, allowing for broad usage and distribution.
Good For
- General Text Generation: Suitable for various text generation tasks where a distilled model with efficient training is beneficial.
- Experimentation: Ideal for developers looking to experiment with models fine-tuned from powerful sources using optimized training techniques.
- Resource-Efficient Deployment: As a Qwen3-based model, it offers a balance of performance and computational efficiency, especially with its GGUF version available for local deployment.