dddsaty/Merge_Sakura_Solar
TEXT GENERATIONConcurrency Cost:1Model Size:10.7BQuant:FP8Ctx Length:4kPublished:Feb 7, 2024License:cc-by-nc-sa-4.0Architecture:Transformer0.0K Open Weights Cold

dddsaty/Merge_Sakura_Solar is a 10.7 billion parameter language model merged from three Sakura-SOLAR-Instruct variants, including one specifically fine-tuned for mathematical tasks. This model leverages the SOLAR architecture and is designed for general instruction-following, with a notable strength in mathematical reasoning. It offers a 4096 token context length and is suitable for applications requiring robust performance across various benchmarks.

Loading preview...