laion/swesmith-316-opt1k__Qwen3-8B
TEXT GENERATIONConcurrency Cost:1Model Size:8BQuant:FP8Ctx Length:32kPublished:Mar 27, 2026License:otherArchitecture:Transformer Cold
The laion/swesmith-316-opt1k__Qwen3-8B is an 8 billion parameter causal language model, fine-tuned from Qwen/Qwen3-8B. This model was trained on the /e/data1/datasets/playground/ot/hf_hub/datasets--laion--swesmith-unified-316/snapshots/2990d3acbbe8e6622cfe408e0f12038e523310ec_thinking_preprocessed dataset, suggesting a specialization in processing or generating content related to the dataset's characteristics. With a 32768 token context length, it is suitable for tasks requiring extensive contextual understanding.
Loading preview...