CharlesLi/llama2_openo1_safe_o1_4o_reflect_4000_1000_full
The CharlesLi/llama2_openo1_safe_o1_4o_reflect_4000_1000_full model is a 7 billion parameter language model fine-tuned from Meta's Llama-2-7b-chat-hf. This model was trained with a focus on safety and reflection, achieving a validation loss of 0.5488. It is intended for applications requiring a Llama-2-based model with enhanced safety characteristics.
Loading preview...
Model Overview
The CharlesLi/llama2_openo1_safe_o1_4o_reflect_4000_1000_full is a 7 billion parameter language model, fine-tuned from the meta-llama/Llama-2-7b-chat-hf base model. The fine-tuning process aimed to enhance the model's safety and reflective capabilities, as indicated by its name and training focus.
Key Training Details
This model was trained using the following hyperparameters:
- Learning Rate: 2e-05
- Batch Size: 4 (train), 4 (eval)
- Gradient Accumulation Steps: 2
- Optimizer: Adam with betas=(0.9, 0.999) and epsilon=1e-08
- LR Scheduler: Cosine with 0.1 warmup ratio
- Epochs: 1
During training, the model achieved a validation loss of 0.5488, with a training loss of 0.7962 at step 100.
Intended Use
Given its fine-tuning from a Llama-2-chat variant and emphasis on "safe" and "reflect" in its naming, this model is likely intended for conversational AI applications where safety and thoughtful responses are prioritized. Developers seeking a Llama-2-based model with these characteristics may find this fine-tune suitable.