CharlesLi/llama_3_gsm8k_per_class_reflect
TEXT GENERATIONConcurrency Cost:1Model Size:8BQuant:FP8Ctx Length:32kPublished:Dec 31, 2024License:llama3.1Architecture:Transformer Cold
The CharlesLi/llama_3_gsm8k_per_class_reflect model is an 8 billion parameter language model fine-tuned from Meta's Llama-3.1-8B-Instruct. It is optimized for tasks related to the GSM8K dataset, likely focusing on mathematical reasoning and problem-solving. This model aims to improve performance on specific arithmetic and logical challenges, achieving a validation loss of 0.5984 during training.
Loading preview...