laion/sft__Kimi-2-5-swesmith-oracle-maxeps-32k__Qwen3-8B
TEXT GENERATIONConcurrency Cost:1Model Size:8BQuant:FP8Ctx Length:32kPublished:Mar 27, 2026License:otherArchitecture:Transformer Cold

This model is a fine-tuned 8 billion parameter Qwen3-8B causal language model, developed by Qwen and further fine-tuned by laion/sft. It is trained with a 32k token context length. The model is fine-tuned on a specific dataset related to 'Kimi-2.5-swesmith-sandboxes-with_tests-oracle_verified_120s-maxeps-32k', suggesting a specialization in tasks related to code or structured problem-solving environments.

Loading preview...