minpeter/calculator-agent-qwen3-0.6b
Hugging Face
TEXT GENERATIONConcurrency Cost:1Model Size:0.8BQuant:BF16Ctx Length:32kPublished:Jan 26, 2026Architecture:Transformer0.0K Warm

The minpeter/calculator-agent-qwen3-0.6b is a 0.8 billion parameter model based on the Qwen3 architecture, fine-tuned for calculator agent tasks. This model, developed by minpeter, demonstrates a significant drop in accuracy compared to its base model, achieving 15.19% on evaluation tasks. It represents an early exploration into reinforcement learning for specific agentic functions, despite its current performance limitations.

Loading preview...

Model Overview

The minpeter/calculator-agent-qwen3-0.6b is a 0.8 billion parameter language model built upon the Qwen3 architecture. This model was developed by minpeter as an initial foray into applying reinforcement learning (RL) for calculator agent tasks.

Key Characteristics

  • Architecture: Qwen3-based.
  • Parameter Count: 0.8 billion parameters.
  • Context Length: 40960 tokens.
  • Training Focus: Fine-tuned for calculator agent functionality using reinforcement learning.

Performance and Limitations

Evaluations indicate that this model currently exhibits a significant performance drop compared to its base model. It achieves an accuracy of 15.19% (24/158) on specific evaluation tasks, whereas the base minpeter/Qwen3-0.6B-Instruct model scored 27.22% (43/158). The creator notes this model as an early, experimental RL project.

When to Consider This Model

Given its current performance, this model is not recommended for production environments where high accuracy in calculator agent tasks is critical. It is primarily suitable for:

  • Research and Development: Exploring early-stage reinforcement learning applications on smaller models.
  • Learning and Experimentation: Understanding the challenges and outcomes of initial RL fine-tuning efforts.
  • Benchmarking: As a baseline for comparing more advanced calculator agent models.