dphn/dolphin-2.8-mistral-7b-v02

TEXT GENERATIONConcurrency Cost:1Model Size:7BQuant:FP8Ctx Length:4kPublished:Mar 28, 2024License:apache-2.0Architecture:Transformer0.2K Open Weights Cold

Dolphin 2.8 Mistral 7b v0.2, developed by Eric Hartford and Cognitive Computations, is a 7 billion parameter language model based on the Mistral-7b-v0.2 architecture. Fine-tuned with a 16k sequence length, it offers a variety of instruction, conversational, and coding skills. This uncensored model is designed for high compliance to requests, making it suitable for developers who implement their own alignment layers.

Loading preview...

Dolphin 2.8 Mistral 7b v0.2 Overview

Dolphin 2.8 Mistral 7b v0.2 is a 7 billion parameter language model developed by Eric Hartford and Cognitive Computations. It is built upon the Mistral-7b-v0.2 base model, which features a 32k context length, and was fine-tuned using 16k sequence lengths. The model was trained over three days on 10x L40S GPUs.

Key Capabilities

  • Instruction Following: Excels at understanding and executing diverse instructions.
  • Conversational AI: Capable of engaging in varied conversational tasks.
  • Coding Skills: Possesses abilities in code generation and related tasks.
  • Uncensored Nature: The model is uncensored and highly compliant to requests, including potentially unethical ones, as its dataset was filtered to remove alignment and bias. Users are advised to implement their own alignment layers.

Performance Metrics

Evaluations show competitive performance across several benchmarks:

  • ARC-Challenge: 0.592 acc
  • GSM8K: 0.478 exact match
  • HellaSwag: 0.638 acc
  • MMLU: 0.612 acc
  • TruthfulQA_MC2: 0.518 acc
  • Winogrande: 0.797 acc

Licensing

Dolphin 2.8 is licensed under Apache 2.0, permitting commercial use. It was trained on data generated from GPT-4, among other models.