dganochenko/llama-3-8b-chat
TEXT GENERATIONConcurrency Cost:1Model Size:8BQuant:FP8Ctx Length:8kPublished:May 16, 2024License:llama3Architecture:Transformer Warm

The dganochenko/llama-3-8b-chat model is an 8 billion parameter instruction-tuned generative text model developed by Meta, part of the Llama 3 family. Optimized for dialogue use cases, it utilizes an optimized transformer architecture with Grouped-Query Attention (GQA) for improved inference scalability. This model excels in assistant-like chat applications and outperforms many open-source chat models on common industry benchmarks, trained on over 15 trillion tokens of publicly available data with an 8k context length.

Loading preview...

Popular Sampler Settings

Top 3 parameter combinations used by Featherless users for this model. Click a tab to see each config.

temperature
top_p
top_k
frequency_penalty
presence_penalty
repetition_penalty
min_p