ewqr2130/mistral-inst-v02-dpo
TEXT GENERATIONConcurrency Cost:1Model Size:7BQuant:FP8Ctx Length:8kPublished:Jan 8, 2024License:mitArchitecture:Transformer Open Weights Cold
The ewqr2130/mistral-inst-v02-dpo is a 7 billion parameter language model developed by ewqr2130, based on the Mistral-inst-v02 architecture. This model has undergone Direct Preference Optimization (DPO) for 6000 epochs, enhancing its instruction-following capabilities. With a context length of 8192 tokens, it is designed for general-purpose conversational and instruction-based tasks.
Loading preview...