EpistemeAI/Fireball-Alpaca-Llama3.1.07-8B-Philos-Math-KTO-beta
TEXT GENERATIONConcurrency Cost:1Model Size:8BQuant:FP8Ctx Length:32kPublished:Sep 12, 2024License:apache-2.0Architecture:Transformer Open Weights Warm

EpistemeAI's Fireball-Alpaca-Llama3.1.07-8B-Philos-Math-KTO-beta is an 8 billion parameter Llama 3.1 based model, fine-tuned using the KTO (Kahneman-Tversky Optimization) method. Developed by EpistemeAI2, this model is optimized for reasoning and mathematical tasks, building upon the Fireball-Alpaca-Llama3.1.07-8B-Philos-Math base. It was trained 2x faster using Unsloth and Huggingface's TRL library, making it suitable for applications requiring efficient mathematical and philosophical reasoning.

Loading preview...