LsTam/stellialm_smallfr_qwen7b_9tplus

Hugging Face
TEXT GENERATIONConcurrency Cost:1Model Size:7.6BQuant:FP8Ctx Length:32kPublished:Dec 16, 2024Architecture:Transformer Warm

LsTam/stellialm_smallfr_qwen7b_9tplus is a 7.6 billion parameter language model, fine-tuned from Qwen/Qwen2.5-7B-Instruct by LsTam. This model is specifically optimized for 9 distinct tasks requiring reasoning and strict output formats, alongside general instruction following in French. It demonstrates high accuracy on these specialized tasks and general French language understanding, aiming to provide efficient performance for specific client needs.

Loading preview...

Model Overview

LsTam/stellialm_smallfr_qwen7b_9tplus is a 7.6 billion parameter model developed by LsTam, fine-tuned from the Qwen/Qwen2.5-7B-Instruct architecture. Its primary focus is on achieving high accuracy across 9 specific tasks that demand strong reasoning capabilities and adherence to strict output formats, in both English and French. The model also features extended training in the French language for general instruction following.

Key Capabilities

  • Specialized Task Performance: Excels in 9 distinct tasks, including answer_reformulation, query_reformulation, summarization, keyword_extraction, fill_in_generation, keyword_update, gqa (General Question Answering), true_false evaluation, and mcq (Multiple Choice Questions).
  • Multilingual Proficiency: Optimized for both English and French, with a particular emphasis on French language instruction data.
  • Reasoning and Format Adherence: Designed for tasks requiring precise reasoning and strict output formatting.
  • Competitive Performance: Internal evaluations suggest its 7B variant approaches GPT-4o performance on its specialized tasks, indicating strong capabilities for its size.

Ideal Use Cases

This model is particularly well-suited for applications requiring:

  • Automated content processing where specific transformations or extractions are needed.
  • Customer support or internal tools that benefit from precise query handling and structured responses.
  • French language applications demanding high-quality instruction following and general understanding.
  • Scenarios where smaller, efficient models can deliver performance comparable to larger, more general-purpose LLMs on targeted tasks.