IlyaGusev/vikhr_nemo_orpo_dostoevsky_12b_slerp
TEXT GENERATIONConcurrency Cost:1Model Size:12BQuant:FP8Ctx Length:32kPublished:Oct 6, 2024Architecture:Transformer0.0K Cold

IlyaGusev/vikhr_nemo_orpo_dostoevsky_12b_slerp is a 12 billion parameter language model created by IlyaGusev, formed by merging vikhr_nemo_orpo_dostoevsky_12b and Vikhr-Nemo-12B-Instruct-R-21-09-24 using the SLERP method. This model leverages the strengths of its constituent models to enhance performance, offering a 32768 token context length. It is designed for general language understanding and generation tasks, benefiting from the combined training of its base components.

Loading preview...