Steven10429/qwen14-2wc1p-eos-3-merge
TEXT GENERATIONConcurrency Cost:1Model Size:14.8BQuant:FP8Ctx Length:32kPublished:Feb 15, 2025License:otherArchitecture:Transformer Cold
Steven10429/qwen14-2wc1p-eos-3-merge is a 14.8 billion parameter language model developed by Steven10429, built upon a Qwen1.5 base architecture. This model features a substantial 131,072 token context length and incorporates specific training iterations focused on improving generation length and controlling the End-Of-Sequence (EOS) token behavior. It is designed for applications requiring robust language understanding and generation with an emphasis on controlled output length.
Loading preview...