jackf857/llama-3-8b-base-margin-dpo-4xh100
TEXT GENERATIONConcurrency Cost:1Model Size:8BQuant:FP8Ctx Length:8kPublished:Apr 2, 2026License:llama3Architecture:Transformer Cold
The jackf857/llama-3-8b-base-margin-dpo-4xh100 is an 8 billion parameter Llama 3 base model, fine-tuned using DPO on the HuggingFaceH4/ultrafeedback_binarized dataset. This model is optimized for improved response quality and alignment through direct preference optimization. It is suitable for general language understanding and generation tasks, building upon its Llama 3 architecture.
Loading preview...