gradientai/Llama-3-8B-Instruct-262k
TEXT GENERATIONConcurrency Cost:1Model Size:8BQuant:FP8Ctx Length:8kPublished:Apr 25, 2024License:llama3Architecture:Transformer0.3K Warm
Gradient's Llama-3-8B-Instruct-262k is an 8 billion parameter instruction-tuned language model based on Meta's Llama 3, specifically engineered for significantly extended context understanding. It expands the original Llama 3 8B's 8k context window to over 160k tokens, demonstrating long-context capabilities with minimal additional training. This model is optimized for tasks requiring deep comprehension and processing of very long inputs, making it suitable for complex analytical and conversational applications.
Loading preview...
Popular Sampler Settings
Top 3 parameter combinations used by Featherless users for this model. Click a tab to see each config.
temperature
top_p
–
top_k
–
frequency_penalty
–
presence_penalty
–
repetition_penalty
–
min_p
–