ByteDance-Seed/cudaLLM-8B

Warm
Public
8B
FP8
32768
License: apache-2.0
Hugging Face
Overview

CudaLLM: High-Performance CUDA Kernel Generation

cudaLLM-8B, developed by ByteDance-Seed, is an 8 billion parameter language model built upon the Qwen3-8B architecture. Its primary function is to generate high-performance and syntactically correct CUDA kernels, making it a specialized tool for GPU parallel programming.

Key Capabilities

  • Specialized CUDA Code Generation: The model is fine-tuned through a two-stage training process to understand and produce complex CUDA code.
  • Performance-Oriented: Training involved a refined dataset where performance-based rewards guided the model, aiming for optimized kernel output.
  • Benchmarked Performance: Evaluated against the KernelBench dataset, demonstrating its proficiency in generating functional CUDA code across various complexity levels.

Training Details

The model was trained using the verl library, leveraging two distinct datasets:

  • SFT Dataset: A high-quality collection of CUDA problem-solution pairs, sourced from models like DeepSeek R1, DeepSeek Coder-7B, and Qwen2-32B.
  • RL Dataset: A refined dataset used for the reinforcement learning stage, providing performance-based feedback.

Intended Use Cases

cudaLLM-8B is ideal for developers working with GPU acceleration. It can serve as:

  • A co-pilot for HPC and CUDA developers to accelerate scientific computing and machine learning.
  • A tool for optimizing existing CUDA kernels.
  • A research platform for AI-driven code generation and optimization.

Limitations

Users should note that while designed for correctness, generated code requires rigorous testing. Security and performance variability on different GPU architectures are also considerations, and the model's specialization means limited performance on general programming or natural language tasks.