zai-org/agentlm-7b
TEXT GENERATIONConcurrency Cost:1Model Size:7BQuant:FP8Ctx Length:4kPublished:Oct 16, 2023Architecture:Transformer0.1K Cold
AgentLM-7B is a 7 billion parameter language model developed by THUDM, fine-tuned using interaction trajectories from multiple agent tasks. This model, based on the Llama-2-chat architecture with a 4096-token context length, is specifically designed to enhance generalized agent capabilities in LLMs. It demonstrates robust generalization on unseen agent tasks while maintaining strong general language abilities.
Loading preview...
AgentLM-7B Overview
AgentLM-7B, developed by THUDM, is a 7 billion parameter language model specifically instruction-tuned to enhance agent capabilities in Large Language Models (LLMs). It is built upon the Llama-2-chat architecture and utilizes a 4096-token context length.
Key Capabilities
- Generalized Agent Abilities: AgentLM-7B is the result of "AgentTuning," a novel approach that instruction-tunes LLMs using interaction trajectories across various agent tasks. This method enables the model to develop robust generalization on agent tasks it has not explicitly encountered during training.
- Maintains General Language Abilities: Despite its specialization in agent tasks, the model is designed to retain strong performance in general language understanding and generation.
- Training Data: The model is produced by mixed training on the custom AgentInstruct dataset and the ShareGPT dataset.
- Conversation Format: It follows the conversation format of Llama-2-chat, with a fixed system prompt: "You are a helpful, respectful and honest assistant."
Good For
- Developing AI Agents: Ideal for applications requiring LLMs to perform complex, multi-step tasks or interact dynamically within environments.
- Research in Agentic AI: Provides a strong baseline for exploring and advancing the capabilities of language models in agent-based scenarios.
- Tasks Requiring Robust Generalization: Suitable for use cases where an agent needs to adapt to new, unseen tasks effectively.