zai-org/webrl-llama-3.1-8b
TEXT GENERATIONConcurrency Cost:1Model Size:8BQuant:FP8Ctx Length:32kPublished:Nov 5, 2024License:otherArchitecture:Transformer0.0K Cold

WebRL-Llama-3.1-8B is an 8 billion parameter open-source model developed by Zhipu AI, based on the Llama-3.1 architecture with a 32768 token context length. This model is specifically fine-tuned for web operations, enabling it to complete tasks on various websites such as OpenStreetMap, Reddit, GitLab, and online store content management systems. It demonstrates significantly enhanced performance in web-based reinforcement learning tasks compared to its base model and other general-purpose LLMs.

Loading preview...