ichi234/exp002_stage2_s2_db_merged
TEXT GENERATIONConcurrency Cost:1Model Size:7.6BQuant:FP8Ctx Length:32kPublished:Mar 1, 2026Architecture:Transformer Cold
The ichi234/exp002_stage2_s2_db_merged model is a 7.6 billion parameter language model based on Qwen2.5-7B-Instruct, fine-tuned for advanced competitive tasks. It is specifically optimized to stabilize output formats and maintain legal action rates for ALFWorld environments, and to improve SQL/answer consistency for DBBench tasks. This model excels at agentic reasoning in structured environments, balancing performance across both ALFWorld's THOUGHT+ACTION format and DBBench's Action: Operation/Answer format. It leverages a multi-stage LoRA fine-tuning process and offline distillation using openai/gpt-oss-120b to achieve its specialized capabilities.
Loading preview...