fzzhang/toten_gsm8k_merged_s
TEXT GENERATIONConcurrency Cost:1Model Size:7BQuant:FP8Ctx Length:4kPublished:Feb 17, 2024License:apache-2.0Architecture:Transformer Open Weights Cold

The fzzhang/toten_gsm8k_merged_s is a 7 billion parameter language model with a 4096 token context length. This model is a merged version, likely optimized for specific tasks, though its primary differentiators and specific training objectives are not detailed in the provided information. It is intended for general language model applications where a 7B parameter model is suitable.

Loading preview...