Niansuh/Llama-3-Groq-70B-Tool-Use
Niansuh/Llama-3-Groq-70B-Tool-Use is a 70 billion parameter causal language model, fine-tuned by Niansuh using Direct Preference Optimization (DPO) on the Llama 3 base model. Optimized for advanced tool use and function calling tasks, it achieves a 90.76% overall accuracy on the Berkeley Function Calling Leaderboard (BFCL), making it the top-performing open-source 70B LLM for these specific applications. This model excels at tasks requiring API interactions and structured data manipulation, with an 8192 token context length.
Loading preview...
Llama-3-Groq-70B-Tool-Use: Advanced Function Calling Model
This model is a 70 billion parameter Llama 3 variant, specifically fine-tuned by Niansuh for advanced tool use and function calling. It leverages a combination of full fine-tuning and Direct Preference Optimization (DPO) to enhance its ability to interact with APIs and manipulate structured data.
Key Capabilities
- Exceptional Tool Use: Designed for scenarios requiring precise function calling and API interaction.
- Leading Performance: Achieves a 90.76% overall accuracy on the Berkeley Function Calling Leaderboard (BFCL), positioning it as the best-performing open-source 70B LLM in this category.
- Optimized Architecture: Built on an optimized transformer architecture, fine-tuned from the Llama 3 70B base model.
- Text-to-Text with Enhanced Tooling: Processes text input to generate text output, with a strong focus on structured tool calls.
Good For
- Research & Development: Ideal for exploring and building applications that require robust tool use and function calling.
- API Interaction: Excels at tasks involving dynamic interaction with external APIs.
- Structured Data Manipulation: Suitable for use cases where the model needs to process and generate structured data.
Important Considerations
While highly capable for tool use, this model is sensitive to sampling configurations; temperature=0.5, top_p=0.65 is a recommended starting point. For general knowledge or open-ended conversational tasks, a general-purpose LLM might be more appropriate. Users should implement safety measures and be aware of potential inaccuracies or biases inherited from the base Llama 3 model.