kevinpro/MetaMathOctopus-MAPO-DPO-13B
TEXT GENERATIONConcurrency Cost:1Model Size:13BQuant:FP8Ctx Length:4kPublished:Feb 26, 2024License:apache-2.0Architecture:Transformer Open Weights Cold

The kevinpro/MetaMathOctopus-MAPO-DPO-13B is a 13 billion parameter language model developed by Shuaijie She et al. (MAPO project) with a 4096-token context length. It is specifically fine-tuned using Multilingual Alignment-as-Preference Optimization (MAPO-DPO) to enhance multilingual reasoning capabilities. This model excels in mathematical and numerical reasoning tasks across various languages, demonstrating strong performance on benchmarks like MSVAMP, MGSM, and MNumGLUESub.

Loading preview...