nbeerbower/llama-3-bophades-v3-8B
TEXT GENERATIONConcurrency Cost:1Model Size:8BQuant:FP8Ctx Length:8kLicense:llama3Architecture:Transformer0.0K Warm

nbeerbower/llama-3-bophades-v3-8B is an 8 billion parameter language model based on Llama-3-8B, fine-tuned using Direct Preference Optimization (DPO). It was trained on a combination of jondurbin/truthy-dpo-v0.1 and kyujinpy/orca_math_dpo datasets, focusing on improving truthfulness and mathematical reasoning. This model is designed for tasks requiring accurate factual responses and robust mathematical problem-solving capabilities.

Loading preview...