xiaolesu/Lean4-sft-tk-8b
TEXT GENERATIONConcurrency Cost:1Model Size:8BQuant:FP8Ctx Length:32kPublished:Mar 16, 2026License:apache-2.0Architecture:Transformer Open Weights Cold

The xiaolesu/Lean4-sft-tk-8b model is an 8 billion parameter language model fine-tuned from Qwen/Qwen3-8B. This model is specifically trained on the xiaolesu/lean4-sft-stmt-tk dataset, indicating an optimization for tasks related to Lean 4 theorem proving or formal verification. It leverages a 32768 token context length and was trained using Axolotl with specific Liger optimizations.

Loading preview...