anakin87/zephyr-7b-alpha-sharded

TEXT GENERATIONConcurrency Cost:1Model Size:7BQuant:FP8Ctx Length:4kPublished:Oct 14, 2023License:mitArchitecture:Transformer0.0K Open Weights Cold

anakin87/zephyr-7b-alpha-sharded is a 7 billion parameter language model, a sharded version of HuggingFaceH4's Zephyr-7B-Alpha, which is fine-tuned from Mistral-7B-v0.1. This model is specifically designed to be easily loaded and run on platforms like Google Colab, particularly with 8-bit quantization. It functions as a helpful assistant, trained using Direct Preference Optimization (DPO) on synthetic datasets, and is intended for educational and research purposes.

Loading preview...

Zephyr 7B Alpha - Sharded: An Accessible Assistant Model

This model, anakin87/zephyr-7b-alpha-sharded, is a sharded version of the original Zephyr 7B Alpha developed by HuggingFaceH4. It is a 7 billion parameter language model built upon mistralai/Mistral-7B-v0.1 and fine-tuned using Direct Preference Optimization (DPO) on a mix of publicly available, synthetic datasets.

Key Characteristics

  • Sharded for Accessibility: The primary differentiator of this specific model variant is its sharded nature, making it easier to load and operate on resource-constrained environments like Google Colab.
  • Optimized for Assistant Tasks: Zephyr models are trained to function as helpful assistants, leveraging DPO to enhance performance on benchmarks like MT Bench.
  • Quantization Friendly: It is recommended to load this model with 8-bit quantization for efficient inference, though 4-bit or half-precision loading is also supported.
  • Research and Educational Focus: Due to the removal of in-built alignment during training, the model may generate problematic text and is explicitly recommended for educational and research use only.

Usage Considerations

  • Colab Compatibility: Ideal for users looking to experiment with a 7B parameter model on platforms with limited GPU memory.
  • Instruction Following: Designed to respond as a helpful chatbot, as demonstrated by its chat template usage.
  • Potential for Unsafe Content: Users should be aware of its potential to generate problematic text when prompted, as alignment was intentionally reduced to boost performance.