CharlesLi/llama_3_alpaca_llama_2
TEXT GENERATIONConcurrency Cost:1Model Size:8BQuant:FP8Ctx Length:32kPublished:Dec 31, 2024License:llama3.1Architecture:Transformer Cold

The CharlesLi/llama_3_alpaca_llama_2 is an 8 billion parameter language model, fine-tuned from Meta's Llama-3.1-8B-Instruct. This model was trained for 30 steps with a learning rate of 0.0002 and achieved a validation loss of 0.8428. It is intended for general language generation tasks, building upon the capabilities of its Llama-3.1 base.

Loading preview...