hkust-nlp/drkernel-8b
TEXT GENERATIONConcurrency Cost:1Model Size:8BQuant:FP8Ctx Length:32kPublished:Feb 5, 2026Architecture:Transformer0.0K Cold

DR.Kernel-8B by hkust-nlp is an 8 billion parameter Qwen3-based model specifically designed for GPU kernel generation and optimization, particularly for Triton kernels. Unlike general-purpose code models, it is trained for iterative refinement with execution feedback from KernelGYM, enabling multi-turn optimization. Its primary use case is in kernel generation research, benchmarking, and agentic code refinement under execution-based reward, focusing on generating and optimizing `ModelNew` kernel implementations from PyTorch reference tasks.

Loading preview...