Niansuh/Llama-3-Groq-70B-Tool-Use

TEXT GENERATIONConcurrency Cost:4Model Size:70BQuant:FP8Ctx Length:8kPublished:Jul 20, 2024License:llama3Architecture:Transformer Cold

Niansuh/Llama-3-Groq-70B-Tool-Use is a 70 billion parameter causal language model, fine-tuned by Niansuh using Direct Preference Optimization (DPO) on the Llama 3 base model. Optimized for advanced tool use and function calling tasks, it achieves a 90.76% overall accuracy on the Berkeley Function Calling Leaderboard (BFCL), making it the top-performing open-source 70B LLM for these specific applications. This model excels at tasks requiring API interactions and structured data manipulation, with an 8192 token context length.

Loading preview...

Llama-3-Groq-70B-Tool-Use: Advanced Function Calling Model

This model is a 70 billion parameter Llama 3 variant, specifically fine-tuned by Niansuh for advanced tool use and function calling. It leverages a combination of full fine-tuning and Direct Preference Optimization (DPO) to enhance its ability to interact with APIs and manipulate structured data.

Key Capabilities

  • Exceptional Tool Use: Designed for scenarios requiring precise function calling and API interaction.
  • Leading Performance: Achieves a 90.76% overall accuracy on the Berkeley Function Calling Leaderboard (BFCL), positioning it as the best-performing open-source 70B LLM in this category.
  • Optimized Architecture: Built on an optimized transformer architecture, fine-tuned from the Llama 3 70B base model.
  • Text-to-Text with Enhanced Tooling: Processes text input to generate text output, with a strong focus on structured tool calls.

Good For

  • Research & Development: Ideal for exploring and building applications that require robust tool use and function calling.
  • API Interaction: Excels at tasks involving dynamic interaction with external APIs.
  • Structured Data Manipulation: Suitable for use cases where the model needs to process and generate structured data.

Important Considerations

While highly capable for tool use, this model is sensitive to sampling configurations; temperature=0.5, top_p=0.65 is a recommended starting point. For general knowledge or open-ended conversational tasks, a general-purpose LLM might be more appropriate. Users should implement safety measures and be aware of potential inaccuracies or biases inherited from the base Llama 3 model.