sbintuitions/sarashina2-70b
TEXT GENERATIONConcurrency Cost:4Model Size:70BQuant:FP8Ctx Length:8kPublished:Aug 6, 2024License:mitArchitecture:Transformer0.0K Open Weights Cold
Sarashina2-70B is a 70 billion parameter causal language model developed by SB Intuitions, built on the Llama2 architecture with a RoPE position type. Trained on 2.1 trillion tokens, including a significant portion of Japanese Common Crawl data and English SlimPajama, it features a 102400-token vocabulary. This model is designed for general language tasks, with a particular emphasis on Japanese language processing, and has an 8192-token context length.
Loading preview...