FinaPolat/llama3_1_8b_dpo-1k_ED_thinking
TEXT GENERATIONConcurrency Cost:1Model Size:8BQuant:FP8Ctx Length:32kPublished:Feb 2, 2026License:apache-2.0Architecture:Transformer Open Weights Cold

The FinaPolat/llama3_1_8b_dpo-1k_ED_thinking model is an 8 billion parameter Llama 3.1 based language model developed by FinaPolat. It was fine-tuned using Unsloth and Huggingface's TRL library, building upon the FinaPolat/llama3_1_8b_thinking_ED model. This model is optimized for efficient training and deployment, offering a 32768 token context length.

Loading preview...