driaforall/Tiny-Agent-a-3B

Hugging Face
TEXT GENERATIONConcurrency Cost:1Model Size:3.1BQuant:BF16Ctx Length:32kPublished:Feb 11, 2025License:qwen-researchArchitecture:Transformer0.0K Warm

Tiny-Agent-a-3B is a 3.1 billion parameter language model developed by driaforall, built upon the Qwen2.5-Coder series. It is specifically fine-tuned for Pythonic function calling, enabling dynamic action creation and complex solution generation through Python code. This model is optimized for edge devices and excels at agentic tasks, outperforming many larger models in Pythonic function calling benchmarks.

Loading preview...

Overview

Tiny-Agent-a-3B is a 3.1 billion parameter model from driaforall, an extension of Dria-Agent-a, specifically designed for agentic tasks on edge devices. It is built on the Qwen2.5-Coder architecture and features quantization-aware training to minimize performance degradation post-quantization.

Key Capabilities

  • Pythonic Function Calling: Unlike traditional JSON-based methods, Tiny-Agent-a-3B uses Python code blocks to interact with tools, allowing for more flexible and powerful agentic behavior. This approach is inspired by works like DynaSaur.
  • One-shot Parallel Multiple Function Calls: The model can execute multiple synchronous processes within a single chat turn, streamlining complex tasks that would otherwise require several conversational exchanges.
  • Free-form Reasoning and Actions: It provides natural language reasoning alongside actions embedded in Python code blocks, mitigating performance loss often associated with rigid output formats.
  • On-the-fly Complex Solution Generation: By generating Python programs (with restricted built-ins), the model can implement custom logic, conditionals, and synchronous pipelines, enabling sophisticated problem-solving beyond what current JSON-based methods typically offer.

Performance

Tiny-Agent-a-3B demonstrates strong performance on the Dria-Pythonic-Agent-Benchmark (DPAB), achieving a score of 72 in Pythonic function calling. This score is competitive with and often surpasses much larger open and closed models, including Qwen-2.5-Coder-32b-Instruct (68) and Llama-3.1-405B-Instruct (60), highlighting its efficiency and effectiveness for agentic workflows.

Good For

  • Edge Device Deployment: Optimized for performance on resource-constrained hardware.
  • Complex Agentic Tasks: Ideal for applications requiring dynamic tool use and sophisticated multi-step reasoning.
  • Python-based Automation: Developers needing an LLM that can generate and execute Python code for interacting with external systems or APIs.