Yaseal/llama3_3b_instruct_vallina_full_sft_30k
TEXT GENERATIONConcurrency Cost:1Model Size:3.2BQuant:BF16Ctx Length:32kPublished:Mar 21, 2026License:otherArchitecture:Transformer Cold
Yaseal/llama3_3b_instruct_vallina_full_sft_30k is a 3.2 billion parameter instruction-tuned language model, fine-tuned from LLM-Research/Llama-3.2-3B-Instruct. It was specifically trained on the deepmath_plain_30k_train dataset, suggesting an optimization for mathematical reasoning and related tasks. This model offers a context length of 32768 tokens, making it suitable for applications requiring processing of extensive input sequences, particularly within specialized domains like mathematics.
Loading preview...