v1olet/v1olet_merged_dpo_7B
TEXT GENERATIONConcurrency Cost:1Model Size:7BQuant:FP8Ctx Length:4kPublished:Dec 12, 2023License:apache-2.0Architecture:Transformer0.0K Open Weights Cold
v1olet/v1olet_merged_dpo_7B is a 7 billion parameter language model developed by Trong-Hieu Nguyen-Mau, fine-tuned using DPO (Direct Preference Optimization). This model is derived from a base model that achieved a top ranking on the 7B leaderboard, indicating strong performance. It is optimized for general language tasks and follows the Alpaca instruction template, making it suitable for various conversational and instruction-following applications.
Loading preview...