universalml/NepaliGPT-2.0
TEXT GENERATIONConcurrency Cost:1Model Size:8BQuant:FP8Ctx Length:32kPublished:Aug 19, 2024License:mitArchitecture:Transformer0.0K Open Weights Cold
NepaliGPT-2.0 is an 8 billion parameter causal language model developed by PrinceLab Pvt. Ltd., fine-tuned from Meta's Llama-3.1-8B-Instruct. This model leverages Unsloth and Huggingface's TRL library for accelerated training, making it a specialized option for applications requiring a Llama-3.1-8B-Instruct base with optimized training efficiency. Its primary use case is for tasks where a Llama-3.1-8B-Instruct derivative is suitable, benefiting from faster fine-tuning.
Loading preview...