anakin87/Llama-3-8b-ita-ties-pro
TEXT GENERATIONConcurrency Cost:1Model Size:8BQuant:FP8Ctx Length:8kPublished:May 18, 2024License:llama3Architecture:Transformer0.0K Warm
anakin87/Llama-3-8b-ita-ties-pro is an 8 billion parameter language model based on the Llama 3 architecture, created by anakin87 using the TIES merge method. It combines two Italian LLMs, DeepMount00/Llama-3-8b-Ita and swap-uniba/LLaMAntino-3-ANITA-8B-Inst-DPO-ITA, with Meta-Llama-3-8B-Instruct as its base. This model is specifically designed and optimized for Italian language tasks, offering a context length of 8192 tokens.
Loading preview...
Popular Sampler Settings
Top 3 parameter combinations used by Featherless users for this model. Click a tab to see each config.
temperature
–
top_p
–
top_k
–
frequency_penalty
–
presence_penalty
–
repetition_penalty
–
min_p
–