DCAgent/a1-stack_go
TEXT GENERATIONConcurrency Cost:1Model Size:8BQuant:FP8Ctx Length:32kPublished:Mar 27, 2026License:otherArchitecture:Transformer Cold

DCAgent/a1-stack_go is an 8 billion parameter language model fine-tuned from Qwen/Qwen3-8B, designed for specific applications related to the 'exp_rpt_stack-go-v3-test_10k_glm_4.7_traces_jupiter' dataset. This model leverages a 32768 token context length and was trained with a focus on specialized data, indicating its potential for tasks within its fine-tuning domain. Its development involved a multi-GPU setup with a cosine learning rate scheduler over 7 epochs.

Loading preview...