ShourenWSR/HT-ht-analysis-Qwen-instruct-no-think-only

TEXT GENERATIONConcurrency Cost:1Model Size:7.6BQuant:FP8Ctx Length:32kPublished:Sep 23, 2025License:otherArchitecture:Transformer Cold

The ShourenWSR/HT-ht-analysis-Qwen-instruct-no-think-only model is a 7.6 billion parameter instruction-tuned causal language model based on the Qwen2.5-7B-Instruct architecture. It has been fine-tuned specifically on the ht-analysis_no_think_only dataset, suggesting an optimization for particular analytical tasks. This model is designed for applications requiring focused analysis without explicit 'thought' processes, leveraging its 32768-token context length for processing substantial inputs.

Loading preview...

Overview

This model, named Qwen_instruct_no_think_only, is a specialized fine-tune of the Qwen/Qwen2.5-7B-Instruct base model. It features 7.6 billion parameters and supports a substantial context length of 32768 tokens, making it suitable for processing extensive textual information.

Key Characteristics

  • Base Model: Qwen2.5-7B-Instruct, a robust causal language model.
  • Fine-tuning: Specifically trained on the ht-analysis_no_think_only dataset.
  • Training Configuration: Utilized a learning rate of 1e-05, a total batch size of 24, and a cosine learning rate scheduler over 3 epochs.

Potential Use Cases

Given its fine-tuning on an 'analysis_no_think_only' dataset, this model is likely optimized for:

  • Direct analytical tasks where explicit step-by-step reasoning or 'thought' processes are not required in the output.
  • Applications demanding concise, direct answers or classifications based on input analysis.
  • Scenarios benefiting from a large context window for comprehensive data processing.