CorticalStack/neurotic-crown-clown-7b-tak-stack-dpo
CorticalStack/neurotic-crown-clown-7b-tak-stack-dpo is a 7 billion parameter language model developed by CorticalStack. It is a DPO fine-tuned version of neurotic-crown-clown-7b-ties, utilizing the CorticalStack/tak-stack-dpo dataset. This model is optimized through Direct Preference Optimization (DPO) to enhance its performance based on specific preferences. With an 8192 token context length, it is suitable for tasks requiring nuanced understanding and generation.
Loading preview...
neurotic-crown-clown-7b-tak-stack-dpo Overview
CorticalStack/neurotic-crown-clown-7b-tak-stack-dpo is a 7 billion parameter language model that has undergone Direct Preference Optimization (DPO) fine-tuning. This model is built upon the base of CorticalStack/neurotic-crown-clown-7b-ties and leverages the specialized CorticalStack/tak-stack-dpo dataset for its DPO training.
Key Training Details
The DPO fine-tuning process involved specific LoRA (Low-Rank Adaptation) configurations and training arguments:
- LoRA Parameters:
r: 32LoRA alpha: 32LoRA dropout: 0.05
- Training Arguments:
Batch size: 4Gradient accumulation steps: 4Optimizer: paged_adamw_32bitMax steps: 100Learning rate: 5e-05Learning rate scheduler type: cosineBeta: 0.1Max prompt length: 1024Max length: 1536
This DPO-tuned model is designed to exhibit improved performance aligned with the preferences encoded in the tak-stack-dpo dataset, making it suitable for applications where such alignment is critical.
Top 3 parameter combinations used by Featherless users for this model. Click a tab to see each config.