mmnga/ELYZA-japanese-Llama-2-7b-fast-instruct-AWQ-calib-ja-100k
TEXT GENERATIONConcurrency Cost:1Model Size:1BQuant:BF16Ctx Length:32kLicense:llama2Architecture:Transformer0.0K Open Weights Cold
The mmnga/ELYZA-japanese-Llama-2-7b-fast-instruct-AWQ-calib-ja-100k model is an AWQ quantized version of ELYZA's 7 billion parameter Llama 2-based instruction-tuned model, specifically optimized for Japanese language tasks. This model was created by mmnga using a Japanese calibration set derived from 100k random samples of Wikipedia data and 200 input/output pairs from ELYZA-tasks-100. Its primary differentiation lies in its AWQ quantization with a Japanese-specific calibration set, aiming to preserve important weights for improved performance in Japanese language generation while reducing model size.
Loading preview...