laion/exp_tas_low_diversity_traces

TEXT GENERATIONConcurrency Cost:1Model Size:8BQuant:FP8Ctx Length:32kPublished:Dec 31, 2025License:apache-2.0Architecture:Transformer Open Weights Cold

The laion/exp_tas_low_diversity_traces model is an 8 billion parameter language model, fine-tuned from Qwen/Qwen3-8B. It was trained on the DCAgent/exp_tas_low_diversity_traces dataset. This model is specifically adapted for tasks related to low-diversity traces, leveraging its Qwen3-8B base for specialized applications.

Loading preview...

Overview

This model, laion/exp_tas_low_diversity_traces, is an 8 billion parameter language model derived from the Qwen/Qwen3-8B architecture. It has undergone specific fine-tuning using the DCAgent/exp_tas_low_diversity_traces dataset.

Key Training Details

The fine-tuning process utilized the following hyperparameters:

  • Learning Rate: 4e-05
  • Batch Size: 1 (train), 8 (eval)
  • Gradient Accumulation: 2 steps, resulting in a total effective batch size of 16
  • Optimizer: ADAMW_TORCH_FUSED with betas=(0.9, 0.98) and epsilon=1e-08
  • Scheduler: Cosine learning rate scheduler with a 0.1 warmup ratio
  • Epochs: 7.0

Intended Use

While specific intended uses and limitations are not detailed in the provided information, its fine-tuning on a dataset focused on "low-diversity traces" suggests potential applications in areas requiring specialized understanding or generation related to such data patterns. Users should be aware that detailed information regarding its specific capabilities, performance, and limitations is currently limited and would require further investigation or documentation from the model's developers.