1024m/QWEN-14B-B100
TEXT GENERATIONConcurrency Cost:1Model Size:14.8BQuant:FP8Ctx Length:32kPublished:Jan 23, 2025License:apache-2.0Architecture:Transformer Open Weights Cold
1024m/QWEN-14B-B100 is a 14.8 billion parameter Qwen2 model developed by 1024m. This model was fine-tuned from unsloth/Qwen2.5-14B-Instruct and optimized for training speed using Unsloth and Huggingface's TRL library. It offers a 32768 token context length, making it suitable for applications requiring efficient processing of longer sequences. Its primary differentiator is its accelerated training methodology, which can be beneficial for developers seeking performant models with optimized development cycles.
Loading preview...