alexneakameni/Qwen2.5-Math-1.5B-Instruct-chess-grpo
Hugging Face
TEXT GENERATIONConcurrency Cost:1Model Size:1.5BQuant:BF16Ctx Length:32kPublished:Dec 30, 2025Architecture:Transformer Warm

The alexneakameni/Qwen2.5-Math-1.5B-Instruct-chess-grpo model is a 1.5 billion parameter Qwen2.5-based instruction-tuned language model developed by alexneakameni. It is specifically fine-tuned for chess-related tasks, designed to output both a rationale and a move in UCI format given a FEN and legal moves. This model excels at generating strategic chess moves and explanations, making it suitable for integrating AI into chess applications or analysis tools.

Loading preview...

Model Overview

The alexneakameni/Qwen2.5-Math-1.5B-Instruct-chess-grpo is a 1.5 billion parameter instruction-tuned model based on the Qwen2.5 architecture, developed by alexneakameni. This model is uniquely specialized for chess, focusing on generating both a rationale and a specific move in Universal Chess Interface (UCI) format.

Key Capabilities

  • Chess Move Generation: Given a FEN (Forsyth-Edwards Notation) and a list of legal moves, the model can propose a move.
  • Strategic Rationale: It provides a short explanation or principal variation (PV) line for its suggested move, enclosed in <rationale>...</rationale> tags.
  • UCI Move Output: The model outputs the chosen move within <uci_move>...</uci_move> tags, facilitating programmatic parsing.
  • Integration: Designed for easy integration with transformers and vLLM for inference, including an OpenAI-compatible endpoint for evaluation.

Good For

  • Chess AI Development: Ideal for developers building chess engines, analysis tools, or AI opponents that require both move suggestions and explanations.
  • Educational Tools: Can be used in applications that teach chess by explaining optimal moves.
  • Research: Provides a specialized base for further research into language models applied to strategic board games like chess.