DCAgent/exp_tas_presence_penalty_1_0_traces

TEXT GENERATIONConcurrency Cost:1Model Size:8BQuant:FP8Ctx Length:32kPublished:Jan 4, 2026License:apache-2.0Architecture:Transformer Open Weights Cold

DCAgent/exp_tas_presence_penalty_1_0_traces is an 8 billion parameter language model fine-tuned from Qwen/Qwen3-8B. This model is specifically adapted using the DCAgent/exp_tas_presence_penalty_1.0_traces dataset. It is designed for tasks related to its fine-tuning data, offering specialized performance within that domain. The model has a context length of 32768 tokens.

Loading preview...

Model Overview

DCAgent/exp_tas_presence_penalty_1_0_traces is an 8 billion parameter language model, fine-tuned from the base Qwen/Qwen3-8B architecture. This specialization is achieved through training on the DCAgent/exp_tas_presence_penalty_1.0_traces dataset, indicating its potential for tasks aligned with the characteristics of this specific data.

Key Training Details

The model was trained with a learning rate of 4e-05 over 7.0 epochs, utilizing a multi-GPU setup with 8 devices. Key hyperparameters included a total training batch size of 16 (with gradient accumulation steps of 2) and an AdamW optimizer with specific beta and epsilon values. A cosine learning rate scheduler with a 0.1 warmup ratio was also employed.

Potential Use Cases

Given its fine-tuning on a specific dataset, this model is likely best suited for:

  • Specialized tasks that align directly with the DCAgent/exp_tas_presence_penalty_1.0_traces dataset's domain.
  • Research and development exploring the impact of presence penalty traces on language model behavior.
  • Applications requiring nuanced understanding or generation within the context of its training data.