asjshiiqqeqweq/Qwen2-7B-Instruct
TEXT GENERATIONConcurrency Cost:1Model Size:7.6BQuant:FP8Ctx Length:32kPublished:Mar 30, 2026License:apache-2.0Architecture:Transformer Open Weights Cold

Qwen2-7B-Instruct is a 7.6 billion parameter instruction-tuned causal language model developed by Qwen, based on the Transformer architecture. It features SwiGLU activation, attention QKV bias, and group query attention, with an improved tokenizer for multiple natural languages and code. This model supports an extended context length of up to 131,072 tokens using YARN and demonstrates strong performance across language understanding, generation, multilingual, coding, mathematics, and reasoning benchmarks, particularly excelling in coding tasks.

Loading preview...