nbeerbower/Lyra-Gutenberg-mistral-nemo-12B

Warm
Public
12B
FP8
32768
License: cc-by-nc-4.0
Hugging Face
Overview

Model Overview

nbeerbower/Lyra-Gutenberg-mistral-nemo-12B is a 12 billion parameter language model derived from the Sao10K/MN-12B-Lyra-v1 base model. It has been further fine-tuned using the jondurbin/gutenberg-dpo-v0.1 dataset, a process conducted over three epochs on an A100 GPU via Google Colab. This fine-tuning approach, similar to ORPO methods, aims to enhance the model's ability to follow instructions and generate high-quality text.

Key Capabilities & Performance

This model demonstrates a general understanding and instruction-following capability, as indicated by its evaluation on the Open LLM Leaderboard. Key performance metrics include:

  • Average Score: 22.57
  • IFEval (0-Shot): 34.95
  • BBH (3-Shot): 36.99
  • MMLU-PRO (5-shot): 29.20

These scores suggest a foundational ability in various reasoning and knowledge-based tasks, with particular strengths in instruction following and multi-task language understanding.

Good For

  • General text generation: Creating coherent and contextually relevant text.
  • Instruction-following tasks: Responding to prompts and directives effectively.
  • Conversational AI: Developing chatbots or interactive agents that require nuanced responses.