AXCXEPT/EZO-Qwen2.5-72B-Instruct
TEXT GENERATIONConcurrency Cost:4Model Size:72.7BQuant:FP8Ctx Length:32kPublished:Sep 20, 2024License:apache-2.0Architecture:Transformer0.0K Open Weights Warm

AXCXEPT/EZO-Qwen2.5-72B-Instruct is a 72.7 billion parameter instruction-tuned causal language model developed by AXCXEPT, based on the Qwen/Qwen2.5-72B-Instruct architecture with a 131,072 token context length. This model has undergone multiple tuning iterations to enhance overall performance, particularly excelling in Japanese language tasks. It achieved a score higher than GPT-4-Turbo on the Japanese MT Bench using GPT-4o as an evaluator, demonstrating strong multilingual capabilities despite its Japanese focus.

Loading preview...

Popular Sampler Settings

Top 3 parameter combinations used by Featherless users for this model. Click a tab to see each config.

temperature
top_p
top_k
frequency_penalty
presence_penalty
repetition_penalty
min_p