rinna/youri-7b
TEXT GENERATIONConcurrency Cost:1Model Size:7BQuant:FP8Ctx Length:4kPublished:Oct 30, 2023License:llama2Architecture:Transformer0.0K Open Weights Warm

rinna/youri-7b is a 7 billion parameter transformer-based causal language model developed by rinna, continually pre-trained from Llama-2-7b on 40 billion tokens of mixed Japanese and English datasets. This model is specifically optimized for Japanese language tasks, significantly improving performance in this domain. It features a 4096-token context length and utilizes the original Llama-2 tokenizer.

Loading preview...