Nicklandshark/Qwen3-1.7B-Wordle-RL
Hugging Face
TEXT GENERATIONConcurrency Cost:1Model Size:2BQuant:BF16Ctx Length:32kPublished:Jan 22, 2026License:apache-2.0Architecture:Transformer Open Weights Warm

Nicklandshark/Qwen3-1.7B-Wordle-RL is a 2 billion parameter language model, fine-tuned using Reinforcement Learning (RL) from the Sedona/Qwen3-1.7B-Wordle-SFT base model. This Qwen3-based model, with a 40960 token context length, is specifically optimized for tasks related to the game Wordle. Its RL fine-tuning aims to enhance performance in Wordle-specific challenges, making it suitable for research and applications in game AI.

Loading preview...

Overview

Nicklandshark/Qwen3-1.7B-Wordle-RL is a 2 billion parameter language model built upon the Qwen3 architecture. It is a Reinforcement Learning (RL) fine-tuned version of the Sedona/Qwen3-1.7B-Wordle-SFT model, indicating a specialized training approach to improve its capabilities in a particular domain.

Key Characteristics

  • Base Model: Derived from Sedona/Qwen3-1.7B-Wordle-SFT.
  • Fine-tuning Method: Utilizes Reinforcement Learning (RL) for specialized optimization.
  • Parameter Count: Features 2 billion parameters.
  • Context Length: Supports a substantial context window of 40960 tokens.

Primary Use Case

This model is specifically designed and fine-tuned for tasks related to the game Wordle. Its RL-based training suggests an emphasis on strategic decision-making or pattern recognition within the Wordle game environment. Developers interested in game AI, particularly for Wordle-like applications, or those exploring the impact of RL fine-tuning on specific game challenges, would find this model relevant.