winglian/Llama-3-8b-64k-PoSE
TEXT GENERATIONConcurrency Cost:1Model Size:8BQuant:FP8Ctx Length:8kPublished:Apr 24, 2024Architecture:Transformer0.1K Warm

winglian/Llama-3-8b-64k-PoSE is an 8 billion parameter Llama 3 model that extends the original 8K context length to 64K using the PoSE (Position-enhanced Sequence Extension) method. This model was further pre-trained on 300 million tokens from the RedPajama V1 dataset, specifically focusing on data between 6K-8K tokens. It is designed for commercial and research use, particularly for tasks requiring a significantly larger context window than the base Llama 3 model.

Loading preview...

Popular Sampler Settings

Top 3 parameter combinations used by Featherless users for this model. Click a tab to see each config.

temperature
top_p
top_k
frequency_penalty
presence_penalty
repetition_penalty
min_p