theprint/ReWiz-Llama-3.1-8B-v2

TEXT GENERATIONConcurrency Cost:1Model Size:8BQuant:FP8Ctx Length:8kPublished:Nov 2, 2024License:apache-2.0Architecture:Transformer0.0K Open Weights Cold

theprint/ReWiz-Llama-3.1-8B-v2 is an 8 billion parameter Llama 3.1-based causal language model developed by theprint. This model was fine-tuned using Unsloth and Huggingface's TRL library, achieving 2x faster training. It is designed for general language tasks, utilizing an Alpaca-style prompt template for optimal performance.

Loading preview...

Model Overview

theprint/ReWiz-Llama-3.1-8B-v2 is an 8 billion parameter language model developed by theprint, fine-tuned from unsloth/meta-llama-3.1-8b-bnb-4bit. This model leverages the Llama 3.1 architecture and was trained with significant efficiency improvements using Unsloth and Huggingface's TRL library, resulting in 2x faster training times.

Key Characteristics

  • Architecture: Llama 3.1-based, 8 billion parameters.
  • Training Efficiency: Utilizes Unsloth for accelerated fine-tuning.
  • Prompt Format: Optimized for an Alpaca-style prompt template to achieve best results.
  • License: Released under the Apache-2.0 license.

Performance Benchmarks

Evaluated on the Open LLM Leaderboard, ReWiz-Llama-3.1-8B-v2 demonstrates a balanced performance across various metrics. Detailed results are available on the Open LLM Leaderboard and its specific evaluation page. Notable scores include:

  • Avg.: 15.68
  • IFEval (0-Shot): 23.73
  • BBH (3-Shot): 23.77
  • MMLU-PRO (5-shot): 25.67

Recommended Use Cases

This model is suitable for general-purpose language generation and understanding tasks where an 8B parameter model is appropriate. Its efficient training process suggests potential for further fine-tuning on specific downstream applications, particularly those benefiting from the Llama 3.1 base and Alpaca-style instruction following.