Ichsan2895/Merak-7B-v5-PROTOTYPE1
TEXT GENERATIONConcurrency Cost:1Model Size:7BQuant:FP8Ctx Length:4kPublished:Dec 10, 2023License:cc-by-nc-sa-4.0Architecture:Transformer Open Weights Cold
Merak-7B-v5-PROTOTYPE1 by Ichsan2895 is a 7 billion parameter large language model specifically fine-tuned for the Indonesian language. Based on the Mistral-7B-OpenOrca architecture, this prototype leverages QLoRA for efficient fine-tuning and DPO-Trainer for RLHF, enabling operation with 16 GB VRAM. Its primary strength lies in processing and generating content in Bahasa Indonesia, making it suitable for Indonesian-centric NLP applications.
Loading preview...