JamyDohrn/LTE-Qwen3-4B-Base
TEXT GENERATIONConcurrency Cost:1Model Size:4BQuant:BF16Ctx Length:32kPublished:Apr 7, 2026License:apache-2.0Architecture:Transformer Open Weights Loading

JamyDohrn/LTE-Qwen3-4B-Base is a 4 billion parameter language model based on the Qwen3 architecture, enhanced with the LTE (Learning from Trial and Error) approach. This model mitigates exploration stagnation in language models by utilizing self-generated errors as hints, without requiring external expert guidance. It is designed to improve both exploitation and exploration during training, enhancing the model's performance upper bound, particularly in reasoning tasks where self-correction is beneficial. The model supports a context length of 32768 tokens.

Loading preview...