dgambettaphd/M_qw306_run0_gen0_WXS_doc1000_synt64_lr1e-04_acm_LANG
TEXT GENERATIONConcurrency Cost:1Model Size:0.8BQuant:BF16Ctx Length:32kPublished:Mar 13, 2026Architecture:Transformer Warm
The dgambettaphd/M_qw306_run0_gen0_WXS_doc1000_synt64_lr1e-04_acm_LANG model is a 0.8 billion parameter language model with a 32768 token context length. This model is automatically generated and pushed to the Hugging Face Hub. Further specific details regarding its architecture, training, and primary differentiators are not provided in the available documentation.
Loading preview...