DiscoResearch/Llama3-DiscoLeo-Instruct-8B-32k-v0.1
TEXT GENERATIONConcurrency Cost:1Model Size:8BQuant:FP8Ctx Length:8kPublished:May 24, 2024License:llama3Architecture:Transformer0.0K Warm

DiscoResearch/Llama3-DiscoLeo-Instruct-8B-32k-v0.1 is an 8 billion parameter instruction-tuned causal language model developed by DiscoResearch and Occiglot, with support from DFKI and hessian.Ai. Derived from Meta's Llama3-8B, it was continuously pretrained on 65 billion high-quality German tokens and further trained on 100 million tokens for a 32k context length. This model is specifically fine-tuned on a German instruction dataset, making it highly effective for German language tasks and long-context applications.

Loading preview...

Popular Sampler Settings

Top 3 parameter combinations used by Featherless users for this model. Click a tab to see each config.

temperature
top_p
top_k
frequency_penalty
presence_penalty
repetition_penalty
min_p