ShenaoZ/0.000001_ablation_iter_2
TEXT GENERATIONConcurrency Cost:1Model Size:7BQuant:FP8Ctx Length:4kPublished:Apr 18, 2024License:mitArchitecture:Transformer Open Weights Cold
ShenaoZ/0.000001_ablation_iter_2 is a 7 billion parameter language model, fine-tuned from ShenaoZ/0.000001_ablation_iter_1. This model was trained for one epoch with a learning rate of 5e-08 and a context length of 4096 tokens. It is an iterative ablation model, suggesting its purpose is likely for research and development in understanding model behavior or specific task performance.
Loading preview...