mlfoundations-dev/open-o1-sft-original-plus-oh-v3.1
TEXT GENERATIONConcurrency Cost:1Model Size:8BQuant:FP8Ctx Length:32kLicense:llama3.1Architecture:Transformer Warm

The mlfoundations-dev/open-o1-sft-original-plus-oh-v3.1 is an 8 billion parameter causal language model, fine-tuned from mlfoundations-dev/oh-dcft-v3.1-gpt-4o-mini. This model is designed for general language understanding and generation tasks, leveraging a 32768 token context length. It was fine-tuned on the mlfoundations-dev/openo1_sft_original dataset, achieving a validation loss of 0.5022. Its primary strength lies in its foundational fine-tuning for diverse language applications.

Loading preview...