eyad-silx/Quasar-2.0-7B-Thinking

TEXT GENERATIONConcurrency Cost:1Model Size:7.6BQuant:FP8Ctx Length:32kArchitecture:Transformer0.0K Cold

Quasar-2.0-7B-Thinking by eyad-silx is a 7.6 billion parameter language model, fine-tuned from the Quasar-2.0-7B base model. This instruction-tuned variant is optimized for reasoning and generating thoughtful responses, particularly in conversational or question-answering contexts. It leverages a 131,072 token context length, making it suitable for processing extensive inputs and generating detailed outputs. The model is designed for applications requiring nuanced understanding and coherent, extended text generation.

Loading preview...

Model Overview

eyad-silx/Quasar-2.0-7B-Thinking is a 7.6 billion parameter language model, fine-tuned from the base eyad-silx/Quasar-2.0-7B model. This iteration focuses on enhancing the model's ability to generate thoughtful and reasoned responses, making it particularly adept at handling complex queries and conversational interactions. It supports a substantial context length of 131,072 tokens, allowing for deep contextual understanding and comprehensive output generation.

Key Capabilities

  • Enhanced Reasoning: Optimized for generating well-considered and logical answers.
  • Extended Context: Processes and generates text based on a 131,072 token context window.
  • Instruction Following: Fine-tuned using SFT (Supervised Fine-Tuning) for improved adherence to instructions.

Training Details

The model was trained using the TRL library, a framework for Transformer Reinforcement Learning, specifically employing Supervised Fine-Tuning (SFT). This process refines the model's behavior to align with desired response patterns, emphasizing coherent and thoughtful output. The training utilized specific versions of key frameworks including TRL 0.15.0.dev0, Transformers 4.49.0.dev0, and Pytorch 2.5.1.