laion/glm46-glaive-code-assistant-sandboxes-maxeps-131k

TEXT GENERATIONConcurrency Cost:1Model Size:8BQuant:FP8Ctx Length:32kPublished:Dec 17, 2025License:apache-2.0Architecture:Transformer Open Weights Cold

The laion/glm46-glaive-code-assistant-sandboxes-maxeps-131k model is an 8 billion parameter language model, fine-tuned from Qwen/Qwen3-8B. It is specifically trained on the DCAgent2/glm46-glaive-code-assistant-sandboxes-maxeps-131k dataset, indicating a specialization in code assistance and sandbox environments. This model is designed for tasks requiring code generation or understanding within specific computational contexts.

Loading preview...

Model Overview

This model, glm46-glaive-code-assistant-sandboxes-maxeps-131k, is an 8 billion parameter language model derived from the Qwen/Qwen3-8B architecture. It has undergone fine-tuning on the DCAgent2/glm46-glaive-code-assistant-sandboxes-maxeps-131k dataset, suggesting a specialized focus on code-related tasks, particularly within sandbox or assistant environments.

Training Details

The model was trained with a learning rate of 4e-05 over 7 epochs, utilizing a total batch size of 16 across 8 GPUs. The optimizer used was ADAMW_TORCH_FUSED with specific beta and epsilon parameters, and a cosine learning rate scheduler with a 0.1 warmup ratio. This configuration indicates a robust training setup aimed at optimizing performance for its intended domain.

Key Characteristics

  • Base Model: Qwen3-8B
  • Fine-tuning Dataset: DCAgent2/glm46-glaive-code-assistant-sandboxes-maxeps-131k
  • Parameter Count: 8 billion

Potential Use Cases

Given its fine-tuning on a code-assistant and sandbox-focused dataset, this model is likely well-suited for:

  • Code generation and completion within constrained environments.
  • Assisting developers with coding tasks.
  • Interpreting and responding to queries related to code execution in sandboxes.