marin-community/marin-8b-base
TEXT GENERATIONConcurrency Cost:1Model Size:8BQuant:FP8Ctx Length:32kPublished:May 15, 2025License:apache-2.0Architecture:Transformer0.0K Open Weights Warm
Marin 8B Base is an 8 billion parameter autoregressive language model developed by the Marin team at Stanford CRFM, built on the Llama architecture with a 32768-token context length. Trained on 12.7 trillion tokens across diverse datasets including Nemotron-CC, Dolma, and specialized Markdownified datasets, it demonstrates strong performance across a range of benchmarks, often outperforming Llama 3.1 8B and other 7-8B open-source models in its class. This base model is designed for broad applications, particularly excelling in reasoning and general language understanding tasks.
Loading preview...