amphora/orpo-5e-8
TEXT GENERATIONConcurrency Cost:1Model Size:7.6BQuant:FP8Ctx Length:32kPublished:Apr 12, 2026License:apache-2.0Architecture:Transformer Open Weights Cold

The amphora/orpo-5e-8 is a 7.6 billion parameter Qwen2 model developed by amphora, fine-tuned from amphora/math-custom-data. This model was trained using Unsloth and Huggingface's TRL library, achieving 2x faster training. It is designed for general language tasks, leveraging its Qwen2 architecture and efficient training methodology.

Loading preview...

Model Overview

The amphora/orpo-5e-8 is a 7.6 billion parameter Qwen2-based language model developed by amphora. It was fine-tuned from the amphora/math-custom-data model, indicating a potential specialization or enhancement related to mathematical or custom data tasks. The training process utilized Unsloth and Huggingface's TRL library, which enabled a significant speedup, achieving 2x faster training.

Key Characteristics

  • Architecture: Based on the Qwen2 model family.
  • Parameter Count: 7.6 billion parameters.
  • Training Efficiency: Leverages Unsloth and Huggingface TRL for 2x faster fine-tuning.
  • Origin: Fine-tuned from amphora/math-custom-data by amphora.

Potential Use Cases

This model is suitable for applications requiring a capable 7.6B parameter language model, especially where the base amphora/math-custom-data model's characteristics are beneficial. Its efficient training suggests it could be a good candidate for further fine-tuning on specific downstream tasks.