X1AOX1A/WorldModel-Sciworld-Qwen2.5-7B

TEXT GENERATIONConcurrency Cost:1Model Size:7.6BQuant:FP8Ctx Length:32kPublished:Dec 9, 2025License:otherArchitecture:Transformer Cold

The X1AOX1A/WorldModel-Sciworld-Qwen2.5-7B is a 7.6 billion parameter language model, fine-tuned from Qwen/Qwen2.5-7B. This model is specifically adapted for tasks related to the 'sciworld_train_with_env_40630' dataset, suggesting an optimization for scientific world modeling or environment interaction. Its development is part of research exploring whether large language models can function as implicit text-based world models. The model's fine-tuning process involved a learning rate of 1e-05 and 5 epochs, indicating a focused adaptation for its specialized domain.

Loading preview...

WorldModel-Sciworld-Qwen2.5-7B: An Implicit Text-based World Model

This model, developed by X1AOX1A, is a fine-tuned version of the Qwen/Qwen2.5-7B base model, featuring 7.6 billion parameters. It is specifically adapted using the sciworld_train_with_env_40630 dataset, indicating a specialization in tasks related to scientific world modeling or environment interaction within a text-based context.

Key Characteristics & Purpose

  • Base Model: Built upon the robust Qwen2.5-7B architecture.
  • Specialized Fine-tuning: Optimized for the sciworld_train_with_env_40630 dataset, suggesting capabilities in understanding and simulating scientific environments or scenarios from text.
  • Research Focus: Part of a broader research initiative titled "From Word to World: Can Large Language Models be Implicit Text-based World Models?" (arXiv:2512.18832). This implies its utility in exploring advanced AI agentic behaviors and understanding how LLMs can internalize and represent complex world dynamics.

Training Details

The model underwent 5 training epochs with a learning rate of 1e-05, utilizing a distributed setup across 4 GPUs with a total batch size of 128. This fine-tuning process aims to imbue the model with specific knowledge and reasoning capabilities pertinent to its target domain.

Potential Use Cases

This model is particularly suited for research and applications requiring an LLM to act as an implicit world model, especially in scientific or environment-interaction contexts. It could be valuable for:

  • Simulating scientific experiments or processes based on textual descriptions.
  • Developing AI agents that can reason about and interact with text-based environments.
  • Exploring the emergent properties of LLMs in representing complex world states and dynamics.