CharlesLi/llama_3_alpaca_cot_simplest
TEXT GENERATIONConcurrency Cost:1Model Size:8BQuant:FP8Ctx Length:32kPublished:Dec 31, 2024License:llama3.1Architecture:Transformer Cold

CharlesLi/llama_3_alpaca_cot_simplest is an 8 billion parameter instruction-tuned causal language model, fine-tuned from Meta's Llama-3.1-8B-Instruct. This model was trained for 30 steps with a learning rate of 0.0002, achieving a final validation loss of 0.8267. It is intended for general instruction-following tasks, building upon the capabilities of its base Llama-3.1 architecture.

Loading preview...