AmberYifan/llama3-8b-full-pretrain-mix-low-tweet-1m-en-sft
TEXT GENERATIONConcurrency Cost:1Model Size:8BQuant:FP8Ctx Length:8kLicense:llama3Architecture:Transformer Cold
The AmberYifan/llama3-8b-full-pretrain-mix-low-tweet-1m-en-sft model is an 8 billion parameter language model, fine-tuned from AmberYifan/llama3-8b-full-pretrain-mix-low-tweet-1m-en. This model is specifically instruction-tuned on the alpaca_en dataset, making it suitable for general-purpose conversational AI and instruction-following tasks. It leverages the Llama 3 architecture and has a context length of 8192 tokens.
Loading preview...