LsTam/stellialm_smallfr_qwen7b_lead

Hugging Face
TEXT GENERATIONConcurrency Cost:1Model Size:7.6BQuant:FP8Ctx Length:32kArchitecture:Transformer Warm

LsTam/stellialm_smallfr_qwen7b_lead is a 7.6 billion parameter causal language model based on Qwen/Qwen2.5-7B-Instruct, fine-tuned specifically for French language tasks. This model excels at reasoning, instruction following, and task completion in French, leveraging a LoRA adapter trained on a dataset of French exercises and instructions. Its primary strength lies in educational applications and scenarios requiring robust French text understanding and generation.

Loading preview...

Model Overview

LsTam/stellialm_smallfr_qwen7b_lead is a 7.6 billion parameter language model derived from Qwen/Qwen2.5-7B-Instruct. It has been fine-tuned using a LoRA adapter, which modified approximately 10% of its parameters, to specialize in French language processing.

Key Capabilities

  • Enhanced French Language Understanding: Optimized for comprehending and generating French text.
  • Instruction Following: Capable of accurately interpreting and executing instructions given in French.
  • Task Completion: Proficient in completing various tasks, particularly those involving French language.
  • Reasoning in French: Demonstrates improved reasoning abilities within a French linguistic context.

Good For

  • Educational Applications: Ideal for tools and platforms focused on learning or teaching in French.
  • French Text Generation: Suitable for generating coherent and contextually relevant French content.
  • Instruction-Based Systems: Effective in scenarios where precise instruction following in French is critical.

Training Details

The model was fine-tuned on a specialized dataset comprising French-language exercises, instructions, and tasks, ensuring its strong performance in these specific areas. While primarily optimized for French, it also exhibits good performance in other languages.