zai-org/agentlm-13b
TEXT GENERATIONConcurrency Cost:1Model Size:13BQuant:FP8Ctx Length:4kPublished:Oct 8, 2023Architecture:Transformer0.0K Cold

AgentLM-13B is a 13 billion parameter instruction-tuned causal language model developed by THUDM. It is specifically fine-tuned using interaction trajectories from multiple agent tasks, enabling robust generalization on unseen agent tasks. This model maintains strong general language abilities while enhancing its capabilities for agentic applications, making it suitable for tasks requiring sequential decision-making and interaction.

Loading preview...

AgentLM-13B: Enhancing LLMs for Agentic Tasks

AgentLM-13B, developed by THUDM, is a 13 billion parameter language model specifically designed to imbue Large Language Models (LLMs) with advanced agent capabilities. This model is a result of AgentTuning, a novel approach that involves instruction-tuning LLMs using interaction trajectories collected across a diverse set of agent tasks. This method allows AgentLM-13B to develop robust generalization abilities for new, unseen agent tasks.

Key Capabilities

  • Agentic Task Performance: Excels in tasks requiring sequential decision-making, planning, and interaction, demonstrating strong performance on various agent benchmarks.
  • Generalization: Exhibits robust generalization to agent tasks not encountered during training, a key differentiator from models trained on static datasets.
  • Maintained Language Abilities: While specialized for agent tasks, AgentLM-13B retains strong general language understanding and generation capabilities.
  • Llama-2-chat Compatibility: Follows the conversation format of Llama-2-chat models, ensuring familiarity and ease of integration for users accustomed to that architecture.

Good For

  • Developing AI Agents: Ideal for researchers and developers building AI agents that need to interact with environments, make decisions, and perform complex, multi-step tasks.
  • Research in Agentic AI: Provides a strong baseline model for exploring and advancing the field of agentic AI and instruction-tuned LLMs.
  • Applications Requiring Robust Task Execution: Suitable for use cases where an LLM needs to reliably execute a series of actions or respond dynamically within a defined operational context.