X1AOX1A/WorldModel-Webshop-Qwen2.5-7B is a 7.6 billion parameter language model fine-tuned from Qwen/Qwen2.5-7B. This model is specifically adapted for tasks related to the 'webshop_train_70790' dataset, suggesting an optimization for web-based interaction or e-commerce environments. It is part of a research initiative exploring whether large language models can function as implicit text-based world models, particularly in agentic reinforcement learning scenarios.
No reviews yet. Be the first to review!