ChocoLlama/Llama-3-ChocoLlama-8B-instruct
TEXT GENERATIONConcurrency Cost:1Model Size:8BQuant:FP8Ctx Length:8kPublished:Jun 7, 2024License:cc-by-nc-4.0Architecture:Transformer0.0K Open Weights Warm

The ChocoLlama/Llama-3-ChocoLlama-8B-instruct is an 8 billion parameter instruction-tuned causal language model developed by Matthieu Meeus and Anthony Rathé. It is a Dutch language-adapted version of Meta's Llama-3-8B, fine-tuned on 32 billion Dutch tokens and further aligned using Supervised Fine-Tuning (SFT) and Direct Preference Optimization (DPO). This model is specifically optimized for Dutch language generation in conversational settings, achieving state-of-the-art performance on Dutch benchmarks within its weight class.

Loading preview...