FelixChao/WizardDolphin-7B

TEXT GENERATIONConcurrency Cost:1Model Size:7BQuant:FP8Ctx Length:4kPublished:Jan 12, 2024License:apache-2.0Architecture:Transformer Open Weights Cold

FelixChao/WizardDolphin-7B is a 7 billion parameter language model created by FelixChao, formed by merging cognitivecomputations/dolphin-2.6-mistral-7b-dpo-laser and WizardLM/WizardMath-7B-V1.1. This model leverages the Mistral-7B-v0.1 base and is optimized for a blend of general instruction following and mathematical reasoning capabilities. It is suitable for applications requiring both conversational fluency and accurate numerical problem-solving.

Loading preview...

WizardDolphin-7B: A Merged 7B Language Model

WizardDolphin-7B is a 7 billion parameter model developed by FelixChao, created through a strategic merge of two distinct models: cognitivecomputations/dolphin-2.6-mistral-7b-dpo-laser and WizardLM/WizardMath-7B-V1.1. This merging process, utilizing the TIES method, aims to combine the strengths of both foundational models.

Key Capabilities

  • Blended Expertise: Integrates the instruction-following and conversational abilities from the Dolphin model with the strong mathematical reasoning and problem-solving skills of the WizardMath model.
  • Mistral-7B Base: Built upon the mistralai/Mistral-7B-v0.1 architecture, providing a robust and efficient foundation.
  • Parameter Efficiency: At 7 billion parameters, it offers a balance between performance and computational resource requirements.

Good For

  • General Instruction Following: Capable of understanding and executing a wide range of user instructions.
  • Mathematical Reasoning: Suitable for tasks involving numerical calculations, logical deduction, and mathematical problem-solving.
  • Hybrid Applications: Ideal for use cases that require both coherent text generation and accurate quantitative analysis, such as educational tools, technical support, or data analysis assistance.