TongSearch/TongSearch-QR-7B

TEXT GENERATIONConcurrency Cost:1Model Size:7.6BQuant:FP8Ctx Length:32kPublished:May 6, 2025License:mitArchitecture:Transformer0.0K Open Weights Cold

TongSearch-QR-7B is a 7.6 billion parameter language model developed by TongSearch, built upon the Qwen2.5-7B-Instruct architecture. This model is designed for specific applications related to its base architecture, offering a substantial parameter count for robust performance. With a context length of 32768 tokens, it is suitable for tasks requiring extensive contextual understanding and generation.

Loading preview...

TongSearch-QR-7B Overview

TongSearch-QR-7B is a 7.6 billion parameter large language model derived from the Qwen2.5-7B-Instruct architecture. This model leverages the foundational capabilities of Qwen2.5-7B-Instruct, indicating a strong base for instruction-following and general language tasks. Its development by TongSearch suggests a focus on particular applications or optimizations built upon this robust open-source foundation.

Key Capabilities

  • Instruction Following: Inherits the instruction-tuned capabilities of its Qwen2.5-7B-Instruct base.
  • Large Context Window: Features a substantial context length of 32768 tokens, enabling it to process and generate long sequences of text, which is beneficial for complex queries or document analysis.
  • 7.6 Billion Parameters: Offers a significant parameter count for nuanced understanding and generation, balancing performance with computational efficiency.

Good For

  • Applications requiring extensive context: The 32768-token context window makes it suitable for tasks like summarizing long documents, detailed question answering over large texts, or maintaining coherence in extended conversations.
  • Developers familiar with Qwen2.5-7B-Instruct: Users already working with or interested in the Qwen2.5 family will find this model a natural extension, potentially with specific enhancements from TongSearch. More technical details can be found in the official GitHub repository.