Model Overview
QuietImpostor/Gemma-3-4b-it-Opusfied is a 4.24 billion parameter full fine-tune of Google's Gemma 3 4B IT. Unlike standard LoRA adapters, this model underwent a native weight fine-tune, updating all parameters to achieve its specialized capabilities. It was trained using the Muon optimizer on a single NVIDIA A100 80GB GPU, enabling effective convergence of dense layers within a 12,288-token context window.
Key Capabilities
- Voice and Style Emulation: Optimized for transferring specific writing styles and emulating distinct conversational voices.
- Long-Context Reasoning: Maintains robust reasoning capabilities over a 12,288-token context window.
- Full Fine-Tune: All model parameters were updated during training, leading to deeper integration of its specialized characteristics.
- Efficient Training: Utilizes the Muon optimizer and Flash Attention 2 for efficient training with a sequence length of 16,384.
Training Details
The model was trained on a curated mix of datasets to balance capability retention and style injection:
crownelius/Opus4.6-No-Reasoning-260x: Provided core voice and conversational style data.crownelius/Opus-4.5-WritingStyle-1000x: Focused on prose and structural writing nuances.crownelius/GLM-5.0-8000x: Contributed to logic, math, and Chain-of-Thought reasoning capabilities.
Ideal Use Cases
This model is particularly well-suited for applications requiring:
- Generating text that adheres to a specific stylistic or tonal requirement.
- Creative writing tasks where voice and prose style are critical.
- Conversational agents needing to adopt a distinct persona or speaking style.
- Tasks benefiting from long-context understanding combined with stylistic control.