giux78/zefiro-7b-sft-qlora-ITA-v0.5
TEXT GENERATIONConcurrency Cost:1Model Size:7BQuant:FP8Ctx Length:8kPublished:Jan 25, 2024License:apache-2.0Architecture:Transformer Open Weights Warm
giux78/zefiro-7b-sft-qlora-ITA-v0.5 is a 7 billion parameter GPT-like model developed by giux78 and funded by Business Operating System. It is a SFT fine-tuned model specifically optimized for the Italian language, based on the Mistral-7B-v0.1 architecture. This model is designed to serve as a foundational base for Italian conversational tasks, leveraging a filtered version of the UltraChat-ITA dataset for its training.
Loading preview...
Popular Sampler Settings
Top 3 parameter combinations used by Featherless users for this model. Click a tab to see each config.
temperature
top_p
top_k
frequency_penalty
presence_penalty
repetition_penalty
min_p