abeiler/AlphaRep

TEXT GENERATIONConcurrency Cost:1Model Size:7BQuant:FP8Ctx Length:4kArchitecture:Transformer Cold

The abeiler/AlphaRep model is a fine-tuned version of Meta's Llama-2-7b-hf, a 7 billion parameter causal language model. This model was trained with a learning rate of 0.0001 over 1 epoch, utilizing Adam optimizer. Further details on its specific capabilities, training dataset, and intended uses are not provided in the available documentation.

Loading preview...

Overview

abeiler/AlphaRep is a fine-tuned language model based on meta-llama/Llama-2-7b-hf. This model leverages the Llama 2 architecture, which is a 7 billion parameter causal language model.

Training Details

The model was trained using the following hyperparameters:

  • Learning Rate: 0.0001
  • Optimizer: Adam with betas=(0.9, 0.999) and epsilon=1e-08
  • Epochs: 1
  • Batch Size: 4 (train), 8 (eval)
  • Frameworks: Transformers 4.33.3, Pytorch 2.0.0, Datasets 2.12.0, Tokenizers 0.13.3

Limitations and Further Information

Currently, detailed information regarding the specific dataset used for fine-tuning, the model's intended uses, and its performance characteristics is not available. Users should exercise caution and conduct their own evaluations before deploying this model for specific applications, as its unique differentiators and optimal use cases are not yet documented.