radlab/pLLama3.2-3B-DPO
TEXT GENERATIONConcurrency Cost:1Model Size:3.2BQuant:BF16Ctx Length:32kPublished:Oct 17, 2024License:llama3.2Architecture:Transformer Warm
radlab/pLLama3.2-3B-DPO is a 3.2 billion parameter language model developed by radlab, fine-tuned and optimized for the Polish language. This model underwent a two-stage training process, including fine-tuning on 650,000 Polish instructions and subsequent DPO (Direct Preference Optimization) on 100,000 examples focused on correct Polish writing. It excels at generating precise and grammatically sound Polish text, making it ideal for applications requiring high-quality Polish language generation and communication.
Loading preview...