Nero10578/Mistral-7B-Sunda-v1.0
TEXT GENERATIONConcurrency Cost:1Model Size:7BQuant:FP8Ctx Length:4kPublished:Dec 22, 2023License:apache-2.0Architecture:Transformer0.0K Open Weights Cold

Nero10578/Mistral-7B-Sunda-v1.0 is a 7 billion parameter Mistral-7B-v0.1 fine-tune, specifically adapted for the Sundanese language. This model was developed to explore the feasibility of teaching a new language to an existing LLM with limited data and QLoRA fine-tuning. It enhances the base Mistral model's capabilities by adding Sundanese language support, making it suitable for applications requiring interaction in Sundanese. The model maintains a 4096 token context length.

Loading preview...