WiroAI/OpenR1-Qwen-7B-French

Loading
Public
7.6B
FP8
131072
License: apache-2.0
Hugging Face
Overview

Overview of WiroAI/OpenR1-Qwen-7B-French

WiroAI/OpenR1-Qwen-7B-French is a 7.6 billion parameter language model developed by WiroAI, based on the Qwen2.5-Instruct architecture. This model was specifically fine-tuned on the WiroAI/dolphin-r1-french dataset for two epochs, with a focus on enhancing French language capabilities. The primary motivation behind its development was to address the limitations of existing models, such as DeepSeek's distilled models, which sometimes default to reasoning in Chinese or English even when prompted in other languages, and to improve performance in relatively low-resource languages like French.

Key Capabilities

  • Enhanced French Reasoning: The model demonstrates improved reasoning processes in French, aiming for clearer and more coherent thought patterns compared to other models that might struggle with non-English/Chinese reasoning.
  • Large Context Window: With a context length of 131072 tokens, it can handle extensive French texts, enabling deeper understanding and generation for complex prompts.
  • Open-Source Contribution: This model is part of an effort to reproduce the R1 project and contribute to the open-source community by providing a specialized French language model.

Good For

  • French Language Applications: Ideal for use cases requiring robust French language understanding, generation, and complex reasoning.
  • Experimental Evaluation: The developers encourage community evaluation to further assess its performance, particularly in scenarios requiring extensive token generation.
  • Research in Low-Resource Languages: Useful for researchers and developers focusing on improving LLM performance in languages with fewer available training resources.