FinaPolat/llama3_1_8b_dpo-1k_ED

TEXT GENERATIONConcurrency Cost:1Model Size:8BQuant:FP8Ctx Length:32kPublished:Jan 21, 2026License:apache-2.0Architecture:Transformer Open Weights Cold

FinaPolat/llama3_1_8b_dpo-1k_ED is an 8 billion parameter Llama 3.1 model developed by FinaPolat, fine-tuned using Huggingface's TRL library. This model was trained twice as fast leveraging the Unsloth framework. It is a fine-tuned iteration of FinaPolat/llama3_1_8b_sft-1k_ED, designed for efficient deployment and performance.

Loading preview...

Model Overview

FinaPolat/llama3_1_8b_dpo-1k_ED is an 8 billion parameter Llama 3.1 model, developed by FinaPolat. This model is a fine-tuned version of FinaPolat/llama3_1_8b_sft-1k_ED, indicating a progression from a supervised fine-tuned base.

Key Characteristics

  • Architecture: Based on the Llama 3.1 family, providing a robust foundation for language tasks.
  • Parameter Count: Features 8 billion parameters, balancing performance with computational efficiency.
  • Efficient Training: Notably, this model was trained twice as fast by utilizing the Unsloth framework in conjunction with Huggingface's TRL library. This suggests an optimization for faster iteration and deployment.

Intended Use

This model is suitable for applications requiring a Llama 3.1-based language model that benefits from efficient training methodologies. Its fine-tuned nature implies it is optimized for specific tasks or domains, building upon its supervised fine-tuned predecessor.