mii-llm/maestrale-chat-v0.4-alpha-sft
TEXT GENERATIONConcurrency Cost:1Model Size:7BQuant:FP8Ctx Length:4kPublished:May 11, 2024License:cc-by-nc-4.0Architecture:Transformer0.0K Open Weights Cold

Maestrale chat v0.4 alpha SFT is a 7 billion parameter language model developed by @efederici and @mferraretto, based on Mistral-7b and continued pre-training for Italian. It is fine-tuned for chat and instruction following, demonstrating improved truthfulness, mathematical, and reasoning capabilities. This model is specifically optimized for Italian language tasks and includes features like agentic behavior and Mermaid mindmap generation.

Loading preview...