ReDiX/Artemide-3.5

Warm
Public
4B
BF16
4096
1
Sep 7, 2024
License: mit
Hugging Face
Overview

Artemide-3.5 Overview

ReDiX/Artemide-3.5 is an instruction-tuned language model built upon the Phi-3.5-mini-instruct architecture, developed by Microsoft. This 3.82 billion parameter model has been specifically fine-tuned by ReDiX using the proprietary ReDiX/DataForge dataset, which comprises a high-quality mixture of Italian and English multi-turn conversations. The model maintains a context length of 4096 tokens.

Key Capabilities & Performance

Artemide-3.5 demonstrates competitive performance, particularly in Italian language understanding. On the OPEN ITA LLM Leaderboard, it achieves an average score of 57.87, outperforming its base model, Phi-3.5-mini-instruct (56.82), and closely matching or exceeding the larger Meta-Llama-3.1-8B-Instruct (56.97) on specific Italian benchmarks like MMLU_IT (60.16 vs 58.43 for Llama-3.1-8B-Instruct). This indicates its strong proficiency in processing and generating content in Italian, while also supporting English.

Ideal Use Cases

  • Multilingual Chatbots: Developing conversational agents that can fluently interact in both Italian and English.
  • Italian Language Processing: Applications requiring robust understanding and generation of Italian text.
  • Resource-Efficient Deployment: Its 3.82 billion parameter size makes it suitable for scenarios where computational resources are a consideration, offering strong performance for its scale.