CharlesLi/llama_2_cot_simplest_code_math_1_full
TEXT GENERATIONConcurrency Cost:1Model Size:7BQuant:FP8Ctx Length:4kPublished:Jan 20, 2025License:llama2Architecture:Transformer Open Weights Cold

The CharlesLi/llama_2_cot_simplest_code_math_1_full model is a 7 billion parameter Llama-2-chat-hf variant, fine-tuned by CharlesLi. This model is based on the Llama 2 architecture and was trained for 1 epoch with a learning rate of 2e-05. It achieved a loss of 0.7902 on the evaluation set, indicating its performance on the specific generator dataset it was fine-tuned on.

Loading preview...