laion/swesmith-3160__Qwen3-8B
TEXT GENERATIONConcurrency Cost:1Model Size:8BQuant:FP8Ctx Length:32kPublished:Mar 26, 2026License:otherArchitecture:Transformer Cold

The laion/swesmith-3160__Qwen3-8B model is an 8 billion parameter language model, fine-tuned from Qwen/Qwen3-8B. It was trained on the laion/swesmith-unified-3160 dataset, featuring a context length of 32768 tokens. This model is a specialized iteration of the Qwen3 architecture, optimized through specific fine-tuning. Its primary application is for tasks benefiting from its particular training data and fine-tuning process.

Loading preview...