bingbangboom/Qwen3006B-transcriber-beta
TEXT GENERATIONConcurrency Cost:1Model Size:0.8BQuant:BF16Ctx Length:32kPublished:Mar 24, 2026License:apache-2.0Architecture:Transformer Open Weights Warm
The bingbangboom/Qwen3006B-transcriber-beta is a Qwen3-0.6B model developed by bingbangboom, specifically fine-tuned as a post-processor for local Automatic Speech Recognition (ASR) outputs. This model was trained using Unsloth and Huggingface's TRL library, enabling faster training. It is designed to process input transcripts, making it suitable for refining ASR results.
Loading preview...
Overview
The bingbangboom/Qwen3006B-transcriber-beta is a specialized Qwen3-0.6B model developed by bingbangboom, primarily designed as a post-processor for local Automatic Speech Recognition (ASR) systems. It was fine-tuned from unsloth/qwen3-0.6b-unsloth-bnb-4bit and leverages Unsloth and Huggingface's TRL library for efficient training.
Key Capabilities
- ASR Post-Processing: Its core function is to refine and process raw transcripts generated by local ASR systems.
- Optimized Training: Benefits from Unsloth's accelerated training methods, indicating potential for efficient deployment.
- Flexible Deployment: Available in various GGUF quantization formats (
Qwen3.5-0.8B.F16.gguf,Qwen3.5-0.8B.Q8_0.gguf,Qwen3.5-0.8B.Q5_K_M.ggu,Qwen3.5-0.8B.Q4_K_M.gguf) and as a Lora merged safetensor, catering to different hardware and performance needs.
Recommended Usage
- Prompt Format: Expects the input transcript directly, e.g.,
{input transcript}for chat or${output}for Handy integration. - Inference Settings: Recommended settings include a low temperature (0.1),
top_kof 10,top_pof 0.95,min_pof 0.05, and arepeat_penaltyof 1.0. - No System Prompt: The model does not require a system prompt. Users should disable model thinking by adding
{%- set enable_thinking = false %}to the Jinja Prompt Template, particularly noted for LMStudio users.