canbingol/gemma3_1B_base-tr-cpt-only_4th_stage_data
TEXT GENERATIONConcurrency Cost:1Model Size:1BQuant:BF16Ctx Length:32kPublished:Apr 10, 2026Architecture:Transformer Loading
The canbingol/gemma3_1B_base-tr-cpt-only_4th_stage_data model is a 1 billion parameter Gemma-3-1B variant developed by canbingol, specifically continued pre-trained on a 50,000-sample subset (samples 150,000–200,000) of a Turkish web corpus. This model isolates the impact of a specific data shard, making it ideal for research into data ordering effects and incremental adaptation in continued pretraining. It is designed for comparative analysis against other CPT models to understand the sensitivity of the model to particular corpus segments.
Loading preview...