MaziyarPanahi/calme-3.1-baguette-3b
MaziyarPanahi/calme-3.1-baguette-3b is a 3.1 billion parameter causal language model, an advanced iteration of Qwen/Qwen2.5-3B. It has been fine-tuned to enhance its capabilities across general domains in both French and English, making it suitable for bilingual applications. With a 32768 token context length, this model is designed for tasks requiring understanding and generation in both languages.
Loading preview...
MaziyarPanahi/calme-3.1-baguette-3b: A Bilingual Qwen2.5-3B Iteration
This model is a 3.1 billion parameter language model, building upon the Qwen/Qwen2.5-3B architecture. It has undergone specific fine-tuning to improve its performance across general tasks in both French and English, making it a versatile option for bilingual applications. The model supports a substantial context length of 32768 tokens.
Key Characteristics:
- Bilingual Capability: Enhanced for general domain tasks in both French and English.
- Base Model: Developed as an advanced iteration of Qwen/Qwen2.5-3B.
- Parameter Count: Features 3.1 billion parameters.
- Context Window: Supports a 32768 token context length.
- Prompt Template: Utilizes the
ChatMLprompt format for structured conversations.
Considerations for Use:
As a relatively small model, users should be aware that its performance may vary with complex prompts and it can be sensitive to hyperparameters. Quantized GGUF versions are available for efficient deployment. Ethical considerations regarding potential biases and limitations, common to all large language models, should be taken into account, recommending safeguards and human oversight in production environments.