openaccess-ai-collective/neft-exp1

TEXT GENERATIONConcurrency Cost:1Model Size:7BQuant:FP8Ctx Length:4kLicense:apache-2.0Architecture:Transformer Open Weights Cold

openaccess-ai-collective/neft-exp1 is a fine-tuned version of Mistral-7B-v0.1, a 7 billion parameter causal language model developed by Mistral AI. This model was trained using Axolotl with specific hyperparameters including a learning rate of 6e-06 and 4 epochs. While specific primary use cases and differentiators are not detailed, its training on the Mistral-7B-v0.1 base suggests general language understanding and generation capabilities.

Loading preview...

Overview

openaccess-ai-collective/neft-exp1 is a fine-tuned variant of the Mistral-7B-v0.1 model, originally developed by Mistral AI. This model was trained using the Axolotl framework, indicating a focus on efficient and customizable fine-tuning processes. While detailed information regarding its specific intended uses, limitations, and the dataset used for fine-tuning is not provided in the current documentation, its foundation on Mistral-7B-v0.1 suggests a strong base for various natural language processing tasks.

Training Details

The model underwent 4 epochs of training with a learning rate of 6e-06, a total batch size of 16 across 8 GPUs, and an Adam optimizer. The training process utilized a cosine learning rate scheduler with 10 warmup steps. The validation loss at the end of training was 1.3731.

Key Capabilities

  • General Language Understanding: Inherits the robust language comprehension abilities of the Mistral-7B-v0.1 base model.
  • Text Generation: Capable of generating coherent and contextually relevant text.

Good for

  • Experimentation with fine-tuned Mistral-7B models.
  • Applications requiring a 7 billion parameter model for general NLP tasks where further domain-specific fine-tuning might be applied.