The rinna/llama-3-youko-8b-instruct is an 8 billion parameter instruction-tuned causal language model developed by rinna, built upon the Llama 3 architecture. It was fine-tuned using supervised fine-tuning (SFT), Chat Vector, and direct preference optimization (DPO) with a diverse dataset including Japanese and English subsets. This model is designed for instruction-following tasks, adopting the Llama-3 chat format and leveraging a 8192 token context length.
No reviews yet. Be the first to review!