rinna/qwq-bakeneko-32b
TEXT GENERATIONConcurrency Cost:2Model Size:32.8BQuant:FP8Ctx Length:32kPublished:Mar 12, 2025License:apache-2.0Architecture:Transformer0.0K Open Weights Warm
rinna/qwq-bakeneko-32b is a 32.8 billion parameter instruction-tuned reasoning model developed by rinna, based on the Qwen2.5 architecture. It is fine-tuned using Chat Vector and Odds Ratio Preference Optimization (ORPO) to deliver superior performance in Japanese language tasks. This model is specifically designed for reasoning applications and adheres to the Qwen/QwQ-32B chat format. It offers a 131072 token context length, making it suitable for complex Japanese language processing.
Loading preview...
Popular Sampler Settings
Top 3 parameter combinations used by Featherless users for this model. Click a tab to see each config.
temperature
–
top_p
–
top_k
–
frequency_penalty
–
presence_penalty
–
repetition_penalty
–
min_p
–