Nike-Hanmatheekuna/llama3-8b-sft-full
TEXT GENERATIONConcurrency Cost:1Model Size:8BQuant:FP8Ctx Length:8kPublished:May 16, 2024License:llama3Architecture:Transformer Cold

The Nike-Hanmatheekuna/llama3-8b-sft-full model is an 8 billion parameter language model fine-tuned from Meta-Llama-3-8B. It was trained with a learning rate of 2e-05 over 3 epochs, utilizing a cosine learning rate scheduler. This model is a specialized version of the Llama 3 architecture, intended for general language tasks following its supervised fine-tuning.

Loading preview...