devanshamin/Qwen2-1.5B-Instruct-Function-Calling-v1

Hugging Face
TEXT GENERATIONConcurrency Cost:1Model Size:1.5BQuant:BF16Ctx Length:32kPublished:Jul 1, 2024License:apache-2.0Architecture:Transformer Open Weights Warm

devanshamin/Qwen2-1.5B-Instruct-Function-Calling-v1 is a 1.5 billion parameter instruction-tuned causal language model, fine-tuned from Qwen/Qwen2-1.5B-Instruct. This model specializes in function calling, having been trained on the devanshamin/gem-viggo-function-calling dataset. It is designed to accurately extract structured information and invoke tools based on user prompts, making it suitable for automation and data extraction tasks.

Loading preview...

Qwen2-1.5B-Instruct-Function-Calling-v1 Overview

This model is a specialized 1.5 billion parameter instruction-tuned variant of the Qwen2-1.5B-Instruct architecture, developed by devanshamin. Its primary distinction lies in its fine-tuning for function calling capabilities, leveraging the devanshamin/gem-viggo-function-calling dataset. This optimization enables the model to understand user intents and generate structured JSON outputs for tool invocation.

Key Capabilities

  • Function Calling: Excels at interpreting natural language prompts to identify and call predefined functions with correctly extracted arguments.
  • Structured Data Extraction: Capable of extracting specific information from text and formatting it into structured JSON, suitable for programmatic use.
  • Flexible Chat Template: Features an updated chat template that supports prompts both with and without tool definitions, enhancing its adaptability.

Good for

  • Automated Workflows: Ideal for integrating with external tools and APIs by translating user requests into function calls.
  • Information Extraction: Useful for tasks requiring the extraction of specific entities or data points from unstructured text into a structured format.
  • Developer Tools: Can serve as a backend for applications that need to parse user commands and execute corresponding actions.