Mathoctopus/Parallel_xRFT_7B: Multilingual Math Reasoning
Mathoctopus/Parallel_xRFT_7B is a 7 billion parameter model from the MathOctopus series, built upon the LLaMA 2 architecture. Developed by Mathoctopus, this model is specifically engineered for advanced multilingual mathematical problem-solving. It was trained using a parallel-training strategy combined with multilingual rejection sampling (xRFT) on the extensive MGSM8KInstruct Dataset, which covers ten distinct languages.
Key Capabilities
- Multilingual Math Problem Solving: Proficient in solving mathematical problems across ten languages, including English, Swahili, Chinese, Bengali, German, Spanish, French, Japanese, Russian, and Thai.
- Enhanced Performance: Demonstrates strong performance on multilingual math benchmarks like MGSM and MSVAMP, often outperforming other models in its class.
- Rejection Sampling (xRFT): Incorporates multilingual rejection sampling during training to refine its mathematical reasoning abilities.
Good For
- Educational Software: Ideal for integration into educational platforms that require accurate math problem-solving in multiple languages.
- Tutoring Systems: Can power intelligent tutoring systems needing to assist users with math in various linguistic contexts.
- Research in Multilingual NLP: Useful for researchers exploring cross-lingual transfer and mathematical reasoning in LLMs.