CharlesLi/llama_2_rlhf_safe_4o_default_500_full
TEXT GENERATIONConcurrency Cost:1Model Size:7BQuant:FP8Ctx Length:4kPublished:Jan 13, 2025License:llama2Architecture:Transformer Open Weights Cold
CharlesLi/llama_2_rlhf_safe_4o_default_500_full is a 7 billion parameter Llama 2-based causal language model fine-tuned from meta-llama/Llama-2-7b-chat-hf. This model is specifically fine-tuned on a generator dataset, demonstrating a training loss of 1.9297. It is intended for applications requiring a Llama 2 variant optimized through RLHF for safety and specific generative tasks.
Loading preview...