bfavro73/qwen2.5-coder-7b-pandas-dpo-aligned

TEXT GENERATIONConcurrency Cost:1Model Size:7.6BQuant:FP8Ctx Length:32kPublished:Mar 24, 2026Architecture:Transformer Cold

bfavro73/qwen2.5-coder-7b-pandas-dpo-aligned is a 7 billion parameter language model fine-tuned from Qwen2.5-Coder-7B-Instruct. This model specializes in Python data analysis, having been fine-tuned using offline DPO with a preference dataset for this specific task. It offers a 128K token context window and is optimized for coding agents, maintaining strong capabilities in mathematics and general competencies.

Loading preview...

Overview

This model, bfavro73/qwen2.5-coder-7b-pandas-dpo-aligned, is a 7 billion parameter language model fine-tuned from Qwen's Qwen2.5-Coder-7B-Instruct. It leverages offline DPO (Direct Preference Optimization) with a preference dataset specifically curated for Python data analysis, making it highly specialized for such tasks. The model balances representational capacity with runtime requirements, offering a robust solution for developers.

Key Capabilities

  • Specialized Code Generation: Enhanced coding capabilities, particularly for Python data analysis.
  • Broad Competencies: Maintains strengths in mathematics and general language understanding.
  • Extended Context Window: Supports a context window of up to 128K tokens, beneficial for complex coding tasks.
  • Real-world Applications: Provides a strong foundation for developing applications like Coding Agents.

Deployment Considerations

  • GGUF Format: Available in GGUF format, with Q4_K_M quantization recommended for systems with at least 8GB of RAM.
  • Memory Requirements: For CPU inference (GGUF), 32 GB RAM is recommended for smoother operation, while 16 GB is a bare minimum. For GPU inference, 10-12 GB VRAM is suggested for 4-bit quantization, and 16 GB VRAM for FP16 or very long contexts.
  • CPU Requirements: A realistic minimum of 8-12 vCPUs is recommended for heavier coding uses with multiple users and long contexts, allowing sufficient threads for the model and application.