CharlesLi/llama2_openo1_safe_o1_4o_reflect_4000_1000_full

TEXT GENERATIONConcurrency Cost:1Model Size:7BQuant:FP8Ctx Length:4kPublished:Mar 28, 2025License:llama2Architecture:Transformer Open Weights Cold

The CharlesLi/llama2_openo1_safe_o1_4o_reflect_4000_1000_full model is a 7 billion parameter language model fine-tuned from Meta's Llama-2-7b-chat-hf. This model was trained with a focus on safety and reflection, achieving a validation loss of 0.5488. It is intended for applications requiring a Llama-2-based model with enhanced safety characteristics.

Loading preview...

Model Overview

The CharlesLi/llama2_openo1_safe_o1_4o_reflect_4000_1000_full is a 7 billion parameter language model, fine-tuned from the meta-llama/Llama-2-7b-chat-hf base model. The fine-tuning process aimed to enhance the model's safety and reflective capabilities, as indicated by its name and training focus.

Key Training Details

This model was trained using the following hyperparameters:

  • Learning Rate: 2e-05
  • Batch Size: 4 (train), 4 (eval)
  • Gradient Accumulation Steps: 2
  • Optimizer: Adam with betas=(0.9, 0.999) and epsilon=1e-08
  • LR Scheduler: Cosine with 0.1 warmup ratio
  • Epochs: 1

During training, the model achieved a validation loss of 0.5488, with a training loss of 0.7962 at step 100.

Intended Use

Given its fine-tuning from a Llama-2-chat variant and emphasis on "safe" and "reflect" in its naming, this model is likely intended for conversational AI applications where safety and thoughtful responses are prioritized. Developers seeking a Llama-2-based model with these characteristics may find this fine-tune suitable.