The astom-M/matsuo-llm-advanced-phase-d is a 7.6 billion parameter language model, fine-tuned from Qwen/Qwen2.5-7B-Instruct, with a context length of 32768 tokens. This model is specifically optimized for agent tasks, leveraging datasets like u-10bei/dbbench_sft_dataset_react_v4, xlangai/spider, birdsql/bird_mini_dev, and the official Phase B ALFWorld v5 dataset. Its primary strength lies in its performance on complex agentic workflows, making it suitable for applications requiring structured reasoning and interaction.
No reviews yet. Be the first to review!