astom-M/matsuo-llm-advanced-phase-d

Cold
Public
7.6B
FP8
32768
Feb 17, 2026
License: apache-2.0
Hugging Face
Overview

Model Overview

The astom-M/matsuo-llm-advanced-phase-d is a 7.6 billion parameter language model, building upon the Qwen/Qwen2.5-7B-Instruct architecture. It has been specifically fine-tuned to excel in agent tasks, distinguishing it from general-purpose LLMs.

Key Capabilities

  • Agent Task Optimization: The model is trained on specialized datasets to enhance its performance in agentic workflows, including those requiring database interaction and environmental reasoning.
  • Diverse Training Data: Fine-tuned using a combination of datasets such as u-10bei/dbbench_sft_dataset_react_v4 (for DBBench format alignment), xlangai/spider, birdsql/bird_mini_dev (for SQL-related tasks), and the official Phase B ALFWorld v5 dataset (for environmental interaction).
  • Compliance: Training adhered to strict compliance guidelines, ensuring no evaluation data was used in training, no LLM was used for data quality filtering, and inference code remained unmodified.
  • vLLM Compatibility: Designed to be compatible with vLLM v0.13.0+ for efficient inference.

Good For

  • Agent-based Applications: Ideal for use cases involving autonomous agents, task automation, and environments requiring structured decision-making.
  • Database Interaction: Strong performance in tasks that involve querying and interacting with databases, as indicated by its training on DBBench and SQL-related datasets.
  • Complex Reasoning: Suited for scenarios demanding multi-step reasoning and interaction within defined environments, such as those found in ALFWorld.