Locutusque/StockQwen-2.5-7B
TEXT GENERATIONConcurrency Cost:1Model Size:7.6BQuant:FP8Ctx Length:32kPublished:Oct 2, 2024Architecture:Transformer0.0K Cold

Locutusque/StockQwen-2.5-7B is a 7.6 billion parameter language model merged from multiple Qwen2.5 variants, including instruction-tuned, coder, and math models. Utilizing the Model Stock merge method, it combines the strengths of its base models to offer enhanced capabilities across general instruction following, code generation, and mathematical reasoning. This model is designed to be a versatile foundation that requires further fine-tuning for specific applications, leveraging a 131072 token context length.

Loading preview...