HuggingFaceH4/mistral-7b-grok

TEXT GENERATIONConcurrency Cost:1Model Size:7BQuant:FP8Ctx Length:4kPublished:Jan 29, 2024License:apache-2.0Architecture:Transformer0.1K Open Weights Cold

HuggingFaceH4/mistral-7b-grok is a 7 billion parameter language model fine-tuned from mistralai/Mistral-7B-v0.1. This model has been aligned using Constitutional AI to emulate the distinctive style of xAI's Grok assistant. It is primarily intended for applications requiring responses with a similar conversational tone and personality to Grok.

Loading preview...

Overview

HuggingFaceH4/mistral-7b-grok is a 7 billion parameter language model derived from the Mistral-7B-v0.1 architecture. Its core distinction lies in its alignment via Constitutional AI, specifically engineered to replicate the unique conversational style of xAI's Grok assistant.

Key Characteristics

  • Base Model: Fine-tuned from mistralai/Mistral-7B-v0.1.
  • Alignment Method: Utilizes Constitutional AI for behavioral shaping.
  • Stylistic Emulation: Designed to mimic the distinctive tone and personality of xAI's Grok.
  • Performance: Achieved a validation loss of 0.9348 during training.

Training Details

The model was trained with a learning rate of 2e-05 over 1 epoch, using an Adam optimizer and a cosine learning rate scheduler. The training involved a total batch size of 256 across 8 GPUs.

Intended Use Cases

This model is best suited for applications where generating text with a conversational style similar to Grok is desired. Developers can leverage it for chatbots, creative content generation, or interactive agents that require a specific, recognizable persona.