gsar78/GreekLlama-1.1B-base
TEXT GENERATIONConcurrency Cost:1Model Size:1.1BQuant:BF16Ctx Length:2kLicense:apache-2.0Architecture:Transformer0.0K Open Weights Gated Cold

gsar78/GreekLlama-1.1B-base is a 1.1 billion parameter base model, based on a custom Llama-like architecture, developed by gsar78. It was pre-trained on a Wikipedia corpus with a 60/40 English to Greek language ratio for 1 billion tokens. This small, experimental model supports both Greek and English languages, primarily intended for research and development purposes.

Loading preview...